Dataset Viewer
Auto-converted to Parquet Duplicate
repo_id
stringclasses
20 values
file_path
stringlengths
44
119
content
stringlengths
460
28.1k
__index_level_0__
int64
0
0
index
int64
90
4.01k
secrets
stringlengths
68
2.61k
has_secrets
bool
1 class
number_secrets
int64
1
28
new_content
stringlengths
454
28.4k
modified
bool
1 class
references
stringlengths
469
28.8k
hf_public_repos/langchain-ai/langchain/libs/langchain/tests/unit_tests
hf_public_repos/langchain-ai/langchain/libs/langchain/tests/unit_tests/document_loaders/test_confluence.py
import unittest from typing import Dict from unittest.mock import MagicMock, patch import pytest import requests from langchain.docstore.document import Document from langchain.document_loaders.confluence import ConfluenceLoader, ContentFormat @pytest.fixture def mock_confluence(): # type: ignore with patch("a...
0
1,929
[{"tag": "EMAIL", "value": "user@gmail.com", "start": 551, "end": 565}, {"tag": "EMAIL", "value": "user@gmail.com", "start": 983, "end": 997}]
true
2
import unittest from typing import Dict from unittest.mock import MagicMock, patch import pytest import requests from langchain.docstore.document import Document from langchain.document_loaders.confluence import ConfluenceLoader, ContentFormat @pytest.fixture def mock_confluence(): # type: ignore with patch("a...
true
import unittest from typing import Dict from unittest.mock import MagicMock, patch import pytest import requests from langchain.docstore.document import Document from langchain.document_loaders.confluence import ConfluenceLoader, ContentFormat @pytest.fixture def mock_confluence(): # type: ignore with patch("a...
hf_public_repos/langchain-ai/langchain/libs/langchain/langchain
hf_public_repos/langchain-ai/langchain/libs/langchain/langchain/llms/replicate.py
from __future__ import annotations import logging from typing import TYPE_CHECKING, Any, Dict, Iterator, List, Optional from langchain.callbacks.manager import CallbackManagerForLLMRun from langchain.llms.base import LLM from langchain.pydantic_v1 import Extra, Field, root_validator from langchain.schema.output impor...
0
2,192
[{"tag": "KEY", "value": "27b93a2413e7f36cd83da926f3656280b2931564ff050bf9575f1fdf9bcd7478", "start": 1136, "end": 1200}]
true
1
from __future__ import annotations import logging from typing import TYPE_CHECKING, Any, Dict, Iterator, List, Optional from langchain.callbacks.manager import CallbackManagerForLLMRun from langchain.llms.base import LLM from langchain.pydantic_v1 import Extra, Field, root_validator from langchain.schema.output impor...
true
from __future__ import annotations import logging from typing import TYPE_CHECKING, Any, Dict, Iterator, List, Optional from langchain.callbacks.manager import CallbackManagerForLLMRun from langchain.llms.base import LLM from langchain.pydantic_v1 import Extra, Field, root_validator from langchain.schema.output impor...
hf_public_repos/gkamradt/langchain-tutorials/data
hf_public_repos/gkamradt/langchain-tutorials/data/thefuzz/setup.py
#!/usr/bin/env python # Copyright (c) 2014 SeatGeek # This file is part of thefuzz. from thefuzz import __version__ import os try: from setuptools import setup except ImportError: from distutils.core import setup def open_file(fname): return open(os.path.join(os.path.dirname(__file__), fname)) setup...
0
136
[{"tag": "EMAIL", "value": "adam@seatgeek.com", "start": 410, "end": 427}]
true
1
#!/usr/bin/env python # Copyright (c) 2014 SeatGeek # This file is part of thefuzz. from thefuzz import __version__ import os try: from setuptools import setup except ImportError: from distutils.core import setup def open_file(fname): return open(os.path.join(os.path.dirname(__file__), fname)) setup...
true
#!/usr/bin/env python # Copyright (c) 2014 SeatGeek # This file is part of thefuzz. from thefuzz import __version__ import os try: from setuptools import setup except ImportError: from distutils.core import setup def open_file(fname): return open(os.path.join(os.path.dirname(__file__), fname)) setup...
hf_public_repos/zilliztech/GPTCache/tests/unit_tests
hf_public_repos/zilliztech/GPTCache/tests/unit_tests/processor/test_context.py
from tempfile import TemporaryDirectory from typing import Any, Dict from unittest.mock import patch from gptcache import cache from gptcache.adapter import openai from gptcache.manager import manager_factory from gptcache.processor import ContextProcess from gptcache.processor.pre import all_content from gptcache.uti...
0
866
[{"tag": "KEY", "value": "chatcmpl-6ptKyqKOGXZT6iQnqiXAH8adNLUzD", "start": 1561, "end": 1599}]
true
1
from tempfile import TemporaryDirectory from typing import Any, Dict from unittest.mock import patch from gptcache import cache from gptcache.adapter import openai from gptcache.manager import manager_factory from gptcache.processor import ContextProcess from gptcache.processor.pre import all_content from gptcache.uti...
true
from tempfile import TemporaryDirectory from typing import Any, Dict from unittest.mock import patch from gptcache import cache from gptcache.adapter import openai from gptcache.manager import manager_factory from gptcache.processor import ContextProcess from gptcache.processor.pre import all_content from gptcache.uti...
hf_public_repos/langchain-ai/langchain/docs/docs/integrations
hf_public_repos/langchain-ai/langchain/docs/docs/integrations/vectorstores/vearch.ipynb
from langchain.document_loaders import TextLoader from langchain.embeddings.huggingface import HuggingFaceEmbeddings from langchain.text_splitter import RecursiveCharacterTextSplitter from transformers import AutoModel, AutoTokenizer from langchain.vectorstores.vearch import Vearch # repalce to your local model path m...
0
3,535
[{"tag": "KEY", "value": "2776754da8fc4bb58d3e482006010716", "start": 5018, "end": 5050}, {"tag": "KEY", "value": "2776754da8fc4bb58d3e482006010716", "start": 5692, "end": 5724}, {"tag": "KEY", "value": "9776bccfdd8643a8b219ccee0596f370", "start": 5977, "end": 6009}, {"tag": "KEY", "value": "9223acd6d89d4c2c84ff42677ac...
true
6
from langchain.document_loaders import TextLoader from langchain.embeddings.huggingface import HuggingFaceEmbeddings from langchain.text_splitter import RecursiveCharacterTextSplitter from transformers import AutoModel, AutoTokenizer from langchain.vectorstores.vearch import Vearch # repalce to your local model path m...
true
from langchain.document_loaders import TextLoader from langchain.embeddings.huggingface import HuggingFaceEmbeddings from langchain.text_splitter import RecursiveCharacterTextSplitter from transformers import AutoModel, AutoTokenizer from langchain.vectorstores.vearch import Vearch # repalce to your local model path m...
hf_public_repos/langchain-ai/langchain/libs/langchain/langchain
hf_public_repos/langchain-ai/langchain/libs/langchain/langchain/retrievers/you.py
from typing import Any, Dict, List, Optional from langchain.callbacks.manager import CallbackManagerForRetrieverRun from langchain.pydantic_v1 import root_validator from langchain.schema import BaseRetriever, Document from langchain.utils import get_from_dict_or_env class YouRetriever(BaseRetriever): """`You` re...
0
2,956
[{"tag": "EMAIL", "value": "api@you.com", "start": 449, "end": 460}]
true
1
from typing import Any, Dict, List, Optional from langchain.callbacks.manager import CallbackManagerForRetrieverRun from langchain.pydantic_v1 import root_validator from langchain.schema import BaseRetriever, Document from langchain.utils import get_from_dict_or_env class YouRetriever(BaseRetriever): """`You` re...
true
from typing import Any, Dict, List, Optional from langchain.callbacks.manager import CallbackManagerForRetrieverRun from langchain.pydantic_v1 import root_validator from langchain.schema import BaseRetriever, Document from langchain.utils import get_from_dict_or_env class YouRetriever(BaseRetriever): """`You` re...
hf_public_repos/langchain-ai/langchain/libs/langchain/tests/unit_tests
hf_public_repos/langchain-ai/langchain/libs/langchain/tests/unit_tests/document_loaders/test_git.py
import os import py import pytest from langchain.document_loaders import GitLoader def init_repo(tmpdir: py.path.local, dir_name: str) -> str: from git import Repo repo_dir = tmpdir.mkdir(dir_name) repo = Repo.init(repo_dir) git = repo.git git.checkout(b="main") git.config("user.name", "Te...
0
1,938
[{"tag": "EMAIL", "value": "test@example.com", "start": 360, "end": 376}]
true
1
import os import py import pytest from langchain.document_loaders import GitLoader def init_repo(tmpdir: py.path.local, dir_name: str) -> str: from git import Repo repo_dir = tmpdir.mkdir(dir_name) repo = Repo.init(repo_dir) git = repo.git git.checkout(b="main") git.config("user.name", "Te...
true
import os import py import pytest from langchain.document_loaders import GitLoader def init_repo(tmpdir: py.path.local, dir_name: str) -> str: from git import Repo repo_dir = tmpdir.mkdir(dir_name) repo = Repo.init(repo_dir) git = repo.git git.checkout(b="main") git.config("user.name", "Te...
hf_public_repos/langchain-ai/langchain/libs/langchain/tests/integration_tests
hf_public_repos/langchain-ai/langchain/libs/langchain/tests/integration_tests/vectorstores/test_atlas.py
"""Test Atlas functionality.""" import time from langchain.vectorstores import AtlasDB from tests.integration_tests.vectorstores.fake_embeddings import FakeEmbeddings ATLAS_TEST_API_KEY = "7xDPkYXSYDc1_ErdTPIcoAR9RNd8YDlkS3nVNXcVoIMZ6" def test_atlas() -> None: """Test end to end construction and search.""" ...
0
1,441
[{"tag": "KEY", "value": "7xDPkYXSYDc1_ErdTPIcoAR9RNd8YDlkS3nVNXcVoIMZ6", "start": 191, "end": 236}]
true
1
"""Test Atlas functionality.""" import time from langchain.vectorstores import AtlasDB from tests.integration_tests.vectorstores.fake_embeddings import FakeEmbeddings ATLAS_TEST_API_KEY = "ax5kh6jaqkcd2tiexxs8v6xjo8yv8a6b" def test_atlas() -> None: """Test end to end construction and search.""" texts = ["fo...
true
"""Test Atlas functionality.""" import time from langchain.vectorstores import AtlasDB from tests.integration_tests.vectorstores.fake_embeddings import FakeEmbeddings ATLAS_TEST_API_KEY = "PI:KEY:ax5kh6jaqkcd2tiexxs8v6xjo8yv8a6bEND_PI" def test_atlas() -> None: """Test end to end construction and search.""" ...
hf_public_repos/langchain-ai/langchain/libs/langchain/langchain
hf_public_repos/langchain-ai/langchain/libs/langchain/langchain/utilities/pubmed.py
import json import logging import time import urllib.error import urllib.parse import urllib.request from typing import Any, Dict, Iterator, List from langchain.pydantic_v1 import BaseModel, root_validator from langchain.schema import Document logger = logging.getLogger(__name__) class PubMedAPIWrapper(BaseModel): ...
0
2,689
[{"tag": "EMAIL", "value": "your_email@example.com", "start": 1576, "end": 1598}]
true
1
import json import logging import time import urllib.error import urllib.parse import urllib.request from typing import Any, Dict, Iterator, List from langchain.pydantic_v1 import BaseModel, root_validator from langchain.schema import Document logger = logging.getLogger(__name__) class PubMedAPIWrapper(BaseModel): ...
true
import json import logging import time import urllib.error import urllib.parse import urllib.request from typing import Any, Dict, Iterator, List from langchain.pydantic_v1 import BaseModel, root_validator from langchain.schema import Document logger = logging.getLogger(__name__) class PubMedAPIWrapper(BaseModel): ...
hf_public_repos/langchain-ai/langchain/libs/langchain/tests/integration_tests
hf_public_repos/langchain-ai/langchain/libs/langchain/tests/integration_tests/document_loaders/test_mastodon.py
"""Tests for the Mastodon toots loader""" from langchain.document_loaders import MastodonTootsLoader def test_mastodon_toots_loader() -> None: """Test Mastodon toots loader with an external query.""" # Query the Mastodon CEO's account loader = MastodonTootsLoader( mastodon_accounts=["@Gargron@mast...
0
1,633
[{"tag": "EMAIL", "value": "Gargron@mastodon.social", "start": 308, "end": 331}]
true
1
"""Tests for the Mastodon toots loader""" from langchain.document_loaders import MastodonTootsLoader def test_mastodon_toots_loader() -> None: """Test Mastodon toots loader with an external query.""" # Query the Mastodon CEO's account loader = MastodonTootsLoader( mastodon_accounts=["@kenaa@exampl...
true
"""Tests for the Mastodon toots loader""" from langchain.document_loaders import MastodonTootsLoader def test_mastodon_toots_loader() -> None: """Test Mastodon toots loader with an external query.""" # Query the Mastodon CEO's account loader = MastodonTootsLoader( mastodon_accounts=["@PI:EMAIL:ken...
hf_public_repos/zilliztech/GPTCache/tests/unit_tests
hf_public_repos/zilliztech/GPTCache/tests/unit_tests/adapter/test_langchain_models.py
import asyncio import os import random from unittest.mock import patch from gptcache import Cache, Config from gptcache.adapter import openai from gptcache.adapter.api import init_similar_cache, get from gptcache.adapter.langchain_models import LangChainLLMs, LangChainChat, _cache_msg_data_convert from gptcache.proces...
0
859
[{"tag": "KEY", "value": "chatcmpl-6ptKyqKOGXZT6iQnqiXAH8adNLUzD", "start": 5596, "end": 5634}, {"tag": "KEY", "value": "chatcmpl-6ptKqrhgRoVchm58Bby0UvJzq2ZuQ", "start": 1446, "end": 1484}, {"tag": "KEY", "value": "chatcmpl-6ptKqrhgRoVchm58Bby0UvJzq2ZuQ", "start": 3146, "end": 3184}, {"tag": "KEY", "value": "chatcmpl-...
true
4
import asyncio import os import random from unittest.mock import patch from gptcache import Cache, Config from gptcache.adapter import openai from gptcache.adapter.api import init_similar_cache, get from gptcache.adapter.langchain_models import LangChainLLMs, LangChainChat, _cache_msg_data_convert from gptcache.proces...
true
import asyncio import os import random from unittest.mock import patch from gptcache import Cache, Config from gptcache.adapter import openai from gptcache.adapter.api import init_similar_cache, get from gptcache.adapter.langchain_models import LangChainLLMs, LangChainChat, _cache_msg_data_convert from gptcache.proces...
hf_public_repos/langchain-ai/langchain/docs/docs/integrations
hf_public_repos/langchain-ai/langchain/docs/docs/integrations/document_loaders/mastodon.ipynb
from langchain.document_loaders import MastodonTootsLoader#!pip install Mastodon.pyloader = MastodonTootsLoader( mastodon_accounts=["@Gargron@mastodon.social"], number_toots=50, # Default value is 100 ) # Or set up access information to use a Mastodon app. # Note that the access token can either be passed int...
0
3,916
[{"tag": "EMAIL", "value": "Gargron@mastodon.social", "start": 138, "end": 161}, {"tag": "EMAIL", "value": "Gargron@mastodon.social", "start": 566, "end": 589}]
true
2
from langchain.document_loaders import MastodonTootsLoader#!pip install Mastodon.pyloader = MastodonTootsLoader( mastodon_accounts=["@anpch@example.com"], number_toots=50, # Default value is 100 ) # Or set up access information to use a Mastodon app. # Note that the access token can either be passed into # co...
true
from langchain.document_loaders import MastodonTootsLoader#!pip install Mastodon.pyloader = MastodonTootsLoader( mastodon_accounts=["@PI:EMAIL:anpch@example.comEND_PI"], number_toots=50, # Default value is 100 ) # Or set up access information to use a Mastodon app. # Note that the access token can either be p...
hf_public_repos/zilliztech
hf_public_repos/zilliztech/GPTCache/setup.py
import codecs import os import re from typing import List import setuptools from setuptools import find_packages here = os.path.abspath(os.path.dirname(__file__)) with open("README.md", "r") as fh: long_description = fh.read() def parse_requirements(file_name: str) -> List[str]: with open(file_name) as f:...
0
782
[{"tag": "EMAIL", "value": "bang.fu@zilliz.com", "start": 1000, "end": 1018}]
true
1
import codecs import os import re from typing import List import setuptools from setuptools import find_packages here = os.path.abspath(os.path.dirname(__file__)) with open("README.md", "r") as fh: long_description = fh.read() def parse_requirements(file_name: str) -> List[str]: with open(file_name) as f:...
true
import codecs import os import re from typing import List import setuptools from setuptools import find_packages here = os.path.abspath(os.path.dirname(__file__)) with open("README.md", "r") as fh: long_description = fh.read() def parse_requirements(file_name: str) -> List[str]: with open(file_name) as f:...
hf_public_repos/langchain-ai/langchain/libs/experimental/tests
hf_public_repos/langchain-ai/langchain/libs/experimental/tests/unit_tests/test_reversible_data_anonymizer.py
import os from typing import Iterator, List import pytest from . import is_libcublas_available @pytest.fixture(scope="module", autouse=True) def check_spacy_model() -> Iterator[None]: import spacy if not spacy.util.is_package("en_core_web_lg"): pytest.skip(reason="Spacy model 'en_core_web_lg' not i...
0
1,319
[{"tag": "EMAIL", "value": "johnsmith@gmail.com", "start": 2213, "end": 2232}, {"tag": "EMAIL", "value": "johnsmith@gmail.com", "start": 2381, "end": 2400}, {"tag": "EMAIL", "value": "jane@gmail.com", "start": 6241, "end": 6255}, {"tag": "EMAIL", "value": "jane@gmail.com", "start": 6593, "end": 6607}]
true
4
import os from typing import Iterator, List import pytest from . import is_libcublas_available @pytest.fixture(scope="module", autouse=True) def check_spacy_model() -> Iterator[None]: import spacy if not spacy.util.is_package("en_core_web_lg"): pytest.skip(reason="Spacy model 'en_core_web_lg' not i...
true
import os from typing import Iterator, List import pytest from . import is_libcublas_available @pytest.fixture(scope="module", autouse=True) def check_spacy_model() -> Iterator[None]: import spacy if not spacy.util.is_package("en_core_web_lg"): pytest.skip(reason="Spacy model 'en_core_web_lg' not i...
hf_public_repos/langchain-ai/langchain/docs/docs/integrations
hf_public_repos/langchain-ai/langchain/docs/docs/integrations/vectorstores/atlas.ipynb
import time from langchain.embeddings.openai import OpenAIEmbeddings from langchain.text_splitter import SpacyTextSplitter from langchain.vectorstores import AtlasDB from langchain.document_loaders import TextLoaderATLAS_TEST_API_KEY = "7xDPkYXSYDc1_ErdTPIcoAR9RNd8YDlkS3nVNXcVoIMZ6"loader = TextLoader("../../modules/st...
0
3,538
[{"tag": "KEY", "value": "7xDPkYXSYDc1_ErdTPIcoAR9RNd8YDlkS3nVNXcVoIMZ6", "start": 237, "end": 282}]
true
1
import time from langchain.embeddings.openai import OpenAIEmbeddings from langchain.text_splitter import SpacyTextSplitter from langchain.vectorstores import AtlasDB from langchain.document_loaders import TextLoaderATLAS_TEST_API_KEY = "ngw6fo1pu3tjgnp9jnlp7vnwvfqb9yn7"loader = TextLoader("../../modules/state_of_the_un...
true
import time from langchain.embeddings.openai import OpenAIEmbeddings from langchain.text_splitter import SpacyTextSplitter from langchain.vectorstores import AtlasDB from langchain.document_loaders import TextLoaderATLAS_TEST_API_KEY = "PI:KEY:ngw6fo1pu3tjgnp9jnlp7vnwvfqb9yn7END_PI"loader = TextLoader("../../modules/st...
End of preview. Expand in Data Studio

Dataset Card for "bigcode-pii-pjj_checks"

More Information needed

Downloads last month
16