Skip to content

[Bug] trust_remote_code is not supported anymore. #8590

@keviddles

Description

@keviddles

What happened?

I am trying to run the Optimizing prompts for a ReAct agent code from Getting Started III on the homepage.

It fails with "trust_remote_code is not supported anymore."

Steps to reproduce

Running:

import dspy
from dspy.datasets import HotPotQA

dspy.configure(lm=dspy.LM("openai/gpt-4o-mini"))

def search_wikipedia(query: str) -> list[str]:
    results = dspy.ColBERTv2(url="http://20.102.90.50:2017/wiki17_abstracts")(query, k=3)
    return [x["text"] for x in results]

trainset = [x.with_inputs('question') for x in HotPotQA(train_seed=2024, train_size=500).train]
react = dspy.ReAct("question -> answer", tools=[search_wikipedia])

tp = dspy.MIPROv2(metric=dspy.evaluate.answer_exact_match, auto="light", num_threads=24)
optimized_react = tp.compile(react, trainset=trainset)

Results in:

`trust_remote_code` is not supported anymore.
Please check that the Hugging Face dataset 'hotpot_qa' isn't based on a loading script and remove `trust_remote_code`.
If the dataset is based on a loading script, please ask the dataset author to remove it and convert it to a standard format like Parquet.
---------------------------------------------------------------------------
RuntimeError                              Traceback (most recent call last)
Cell In[17], line 10
      7     results = dspy.ColBERTv2(url="http://20.102.90.50:2017/wiki17_abstracts")(query, k=3)
      8     return [x["text"] for x in results]
---> 10 trainset = [x.with_inputs('question') for x in HotPotQA(train_seed=2024, train_size=500).train]
     11 react = dspy.ReAct("question -> answer", tools=[search_wikipedia])
     13 tp = dspy.MIPROv2(metric=dspy.evaluate.answer_exact_match, auto="light", num_threads=24)

File [~/.cache/uv/archive-v0/KO8BfiDHTjQ5IO2ZclrNP/lib/python3.12/site-packages/dspy/datasets/hotpotqa.py:23](http://localhost:8898/lab/tree/notebooks/~/.cache/uv/archive-v0/KO8BfiDHTjQ5IO2ZclrNP/lib/python3.12/site-packages/dspy/datasets/hotpotqa.py#line=22), in HotPotQA.__init__(self, only_hard_examples, keep_details, unofficial_dev, *args, **kwargs)
     16 assert only_hard_examples, (
     17     "Care must be taken when adding support for easy examples."
     18     "Dev must be all hard to match official dev, but training can be flexible."
     19 )
     21 from datasets import load_dataset
---> 23 hf_official_train = load_dataset("hotpot_qa", "fullwiki", split="train", trust_remote_code=True)
     24 hf_official_dev = load_dataset("hotpot_qa", "fullwiki", split="validation", trust_remote_code=True)
     26 official_train = []

File [~/.cache/uv/archive-v0/KO8BfiDHTjQ5IO2ZclrNP/lib/python3.12/site-packages/datasets/load.py:1392](http://localhost:8898/lab/tree/notebooks/~/.cache/uv/archive-v0/KO8BfiDHTjQ5IO2ZclrNP/lib/python3.12/site-packages/datasets/load.py#line=1391), in load_dataset(path, name, data_dir, data_files, split, cache_dir, features, download_config, download_mode, verification_mode, keep_in_memory, save_infos, revision, token, streaming, num_proc, storage_options, **config_kwargs)
   1387 verification_mode = VerificationMode(
   1388     (verification_mode or VerificationMode.BASIC_CHECKS) if not save_infos else VerificationMode.ALL_CHECKS
   1389 )
   1391 # Create a dataset builder
-> 1392 builder_instance = load_dataset_builder(
   1393     path=path,
   1394     name=name,
   1395     data_dir=data_dir,
   1396     data_files=data_files,
   1397     cache_dir=cache_dir,
   1398     features=features,
   1399     download_config=download_config,
   1400     download_mode=download_mode,
   1401     revision=revision,
   1402     token=token,
   1403     storage_options=storage_options,
   1404     **config_kwargs,
   1405 )
   1407 # Return iterable dataset in case of streaming
   1408 if streaming:

File [~/.cache/uv/archive-v0/KO8BfiDHTjQ5IO2ZclrNP/lib/python3.12/site-packages/datasets/load.py:1132](http://localhost:8898/lab/tree/notebooks/~/.cache/uv/archive-v0/KO8BfiDHTjQ5IO2ZclrNP/lib/python3.12/site-packages/datasets/load.py#line=1131), in load_dataset_builder(path, name, data_dir, data_files, cache_dir, features, download_config, download_mode, revision, token, storage_options, **config_kwargs)
   1130 if features is not None:
   1131     features = _fix_for_backward_compatible_features(features)
-> 1132 dataset_module = dataset_module_factory(
   1133     path,
   1134     revision=revision,
   1135     download_config=download_config,
   1136     download_mode=download_mode,
   1137     data_dir=data_dir,
   1138     data_files=data_files,
   1139     cache_dir=cache_dir,
   1140 )
   1141 # Get dataset builder class
   1142 builder_kwargs = dataset_module.builder_kwargs

File [~/.cache/uv/archive-v0/KO8BfiDHTjQ5IO2ZclrNP/lib/python3.12/site-packages/datasets/load.py:1031](http://localhost:8898/lab/tree/notebooks/~/.cache/uv/archive-v0/KO8BfiDHTjQ5IO2ZclrNP/lib/python3.12/site-packages/datasets/load.py#line=1030), in dataset_module_factory(path, revision, download_config, download_mode, data_dir, data_files, cache_dir, **download_kwargs)
   1026             if isinstance(e1, FileNotFoundError):
   1027                 raise FileNotFoundError(
   1028                     f"Couldn't find any data file at {relative_to_absolute_path(path)}. "
   1029                     f"Couldn't find '{path}' on the Hugging Face Hub either: {type(e1).__name__}: {e1}"
   1030                 ) from None
-> 1031             raise e1 from None
   1032 else:
   1033     raise FileNotFoundError(f"Couldn't find any data file at {relative_to_absolute_path(path)}.")

File [~/.cache/uv/archive-v0/KO8BfiDHTjQ5IO2ZclrNP/lib/python3.12/site-packages/datasets/load.py:989](http://localhost:8898/lab/tree/notebooks/~/.cache/uv/archive-v0/KO8BfiDHTjQ5IO2ZclrNP/lib/python3.12/site-packages/datasets/load.py#line=988), in dataset_module_factory(path, revision, download_config, download_mode, data_dir, data_files, cache_dir, **download_kwargs)
    981 try:
    982     api.hf_hub_download(
    983         repo_id=path,
    984         filename=filename,
   (...)    987         proxies=download_config.proxies,
    988     )
--> 989     raise RuntimeError(f"Dataset scripts are no longer supported, but found {filename}")
    990 except EntryNotFoundError:
    991     # Use the infos from the parquet export except in some cases:
    992     if data_dir or data_files or (revision and revision != "main"):

RuntimeError: Dataset scripts are no longer supported, but found hotpot_qa.py

DSPy version

2.6.27

Metadata

Metadata

Assignees

No one assigned

    Labels

    bugSomething isn't working

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions