Loading...
Loading...
Use this skill for Hugging Face Dataset Viewer API workflows that fetch subset/split metadata, paginate rows, search text, apply filters, download parquet URLs, and read size or statistics.
npx skill4agent add huggingface/skills hugging-face-dataset-viewer/is-validconfigsplit/splits/first-rows/rowsoffsetlength/search/filter/parquet/size/statisticshttps://datasets-server.huggingface.coGEToffsetlength100Authorization: Bearer <HF_TOKEN>Validate dataset/is-valid?dataset=<namespace/repo>List subsets and splits/splits?dataset=<namespace/repo>Preview first rows/first-rows?dataset=<namespace/repo>&config=<config>&split=<split>Paginate rows/rows?dataset=<namespace/repo>&config=<config>&split=<split>&offset=<int>&length=<int>Search text/search?dataset=<namespace/repo>&config=<config>&split=<split>&query=<text>&offset=<int>&length=<int>Filter with predicates/filter?dataset=<namespace/repo>&config=<config>&split=<split>&where=<predicate>&orderby=<sort>&offset=<int>&length=<int>List parquet shards/parquet?dataset=<namespace/repo>Get size totals/size?dataset=<namespace/repo>Get column statistics/statistics?dataset=<namespace/repo>&config=<config>&split=<split>Get Croissant metadata (if available)/croissant?dataset=<namespace/repo>curl "https://datasets-server.huggingface.co/rows?dataset=stanfordnlp/imdb&config=plain_text&split=train&offset=0&length=100"
curl "https://datasets-server.huggingface.co/rows?dataset=stanfordnlp/imdb&config=plain_text&split=train&offset=100&length=100"num_rows_totalnum_rows_per_pagepartial/search/filterwhereorderbynpx parquetlenshf://datasets/<namespace>/<repo>@~parquet/<config>/<split>/<shard>.parquet<config><split><shard>/parquetcurl -s "https://datasets-server.huggingface.co/parquet?dataset=cfahlgren1/hub-stats" \
| jq -r '.parquet_files[] | "hf://datasets/\(.dataset)@~parquet/\(.config)/\(.split)/\(.filename)"'npx -y -p parquetlens -p @parquetlens/sql parquetlens \
"hf://datasets/<namespace>/<repo>@~parquet/<config>/<split>/<shard>.parquet" \
--sql "SELECT * FROM data LIMIT 20"--sql "COPY (SELECT * FROM data LIMIT 1000) TO 'export.csv' (FORMAT CSV, HEADER, DELIMITER ',')"--sql "COPY (SELECT * FROM data LIMIT 1000) TO 'export.json' (FORMAT JSON)"--sql "COPY (SELECT * FROM data LIMIT 1000) TO 'export.parquet' (FORMAT PARQUET)"https://huggingface.co/new-datasetcurl -s "https://datasets-server.huggingface.co/parquet?dataset=<namespace>/<repo>"npx @huggingface/hubhfjsexport HF_TOKEN=<your_hf_token>npx -y @huggingface/hub upload datasets/<namespace>/<repo> ./local/parquet-folder datanpx -y @huggingface/hub upload datasets/<namespace>/<repo> ./local/parquet-folder data --private/parquet<config>/<split>/<shard>@~parquet