summaryrefslogtreecommitdiff
path: root/dev-python/pyarrow/files
diff options
context:
space:
mode:
authorV3n3RiX <venerix@koprulu.sector>2023-07-24 14:55:44 +0100
committerV3n3RiX <venerix@koprulu.sector>2023-07-24 14:55:44 +0100
commitb83b16e0f6c963583f6323fe1bb12ecbd11d6f81 (patch)
treed67748047921162af2742c3e1255e3f1421f769b /dev-python/pyarrow/files
parent76617568771869151cb2a3f71939719c15dd47a2 (diff)
gentoo auto-resync : 24:07:2023 - 14:55:44
Diffstat (limited to 'dev-python/pyarrow/files')
-rw-r--r--dev-python/pyarrow/files/pyarrow-12.0.0-tests.patch144
1 files changed, 0 insertions, 144 deletions
diff --git a/dev-python/pyarrow/files/pyarrow-12.0.0-tests.patch b/dev-python/pyarrow/files/pyarrow-12.0.0-tests.patch
deleted file mode 100644
index 56a307592593..000000000000
--- a/dev-python/pyarrow/files/pyarrow-12.0.0-tests.patch
+++ /dev/null
@@ -1,144 +0,0 @@
---- a/pyarrow/tests/test_compute.py 2023-05-08 09:06:34.571387618 +0200
-+++ b/pyarrow/tests/test_compute.py 2023-05-08 09:11:52.759753459 +0200
-@@ -414,6 +414,7 @@
- assert pc.variance(data, ddof=1).as_py() == 6.0
-
-
-+@pytest.mark.skip(reason="not working")
- def test_count_substring():
- for (ty, offset) in [(pa.string(), pa.int32()),
- (pa.large_string(), pa.int64())]:
-@@ -428,6 +429,7 @@
- assert expected == result
-
-
-+@pytest.mark.skip(reason="not working")
- def test_count_substring_regex():
- for (ty, offset) in [(pa.string(), pa.int32()),
- (pa.large_string(), pa.int64())]:
-@@ -442,6 +444,7 @@
- assert expected.equals(result)
-
-
-+@pytest.mark.skip(reason="not working")
- def test_find_substring():
- for ty in [pa.string(), pa.binary(), pa.large_string(), pa.large_binary()]:
- arr = pa.array(["ab", "cab", "ba", None], type=ty)
-@@ -459,6 +462,7 @@
- assert result.to_pylist() == [0, 1, 0, 0]
-
-
-+@pytest.mark.skip(reason="not working")
- def test_match_like():
- arr = pa.array(["ab", "ba%", "ba", "ca%d", None])
- result = pc.match_like(arr, r"_a\%%")
-@@ -474,6 +478,7 @@
- assert expected.equals(result)
-
-
-+@pytest.mark.skip(reason="not working")
- def test_match_substring():
- arr = pa.array(["ab", "abc", "ba", None])
- result = pc.match_substring(arr, "ab")
-@@ -489,6 +494,7 @@
- assert expected.equals(result)
-
-
-+@pytest.mark.skip(reason="not working")
- def test_match_substring_regex():
- arr = pa.array(["ab", "abc", "ba", "c", None])
- result = pc.match_substring_regex(arr, "^a?b")
-@@ -602,6 +608,7 @@
- assert expected.equals(result)
-
-
-+@pytest.mark.skip(reason="not working")
- def test_split_pattern_regex():
- arr = pa.array(["-foo---bar--", "---foo---b"])
- result = pc.split_pattern_regex(arr, pattern="-+")
-@@ -1022,6 +1029,7 @@
- assert ar.tolist() == ['barzfoo', 'bard', None]
-
-
-+@pytest.mark.skip(reason="not working")
- def test_replace_regex():
- data = pa.array(['foo', 'mood', None])
- expected = ['f00', 'm00d', None]
-@@ -1033,6 +1041,7 @@
- assert ar.tolist() == expected
-
-
-+@pytest.mark.skip(reason="not working")
- def test_extract_regex():
- ar = pa.array(['a1', 'zb2z'])
- expected = [{'letter': 'a', 'digit': '1'}, {'letter': 'b', 'digit': '2'}]
---- a/pyarrow/tests/test_fs.py 2023-05-08 09:13:26.796384297 +0200
-+++ b/pyarrow/tests/test_fs.py 2023-05-08 09:14:20.567601499 +0200
-@@ -1012,6 +1012,7 @@
- LocalFileSystem(xxx=False)
-
-
-+@pytest.mark.skip(reason="not working")
- def test_localfs_errors(localfs):
- # Local filesystem errors should raise the right Python exceptions
- # (e.g. FileNotFoundError)
---- a/pyarrow/tests/test_memory.py 2023-05-08 09:15:35.366512597 +0200
-+++ b/pyarrow/tests/test_memory.py 2023-05-08 09:16:44.969501524 +0200
-@@ -140,6 +140,7 @@
- assert len(errlines) == 0
-
-
-+@pytest.mark.skip(reason="not working")
- def test_env_var():
- check_env_var("system", ["system"])
- if should_have_jemalloc:
-@@ -149,6 +150,7 @@
- check_env_var("nonexistent", possible_backends, expect_warning=True)
-
-
-+@pytest.mark.skip(reason="not working")
- def test_specific_memory_pools():
- specific_pools = set()
-
-@@ -170,6 +172,7 @@
- can_fail=not should_have_mimalloc)
-
-
-+@pytest.mark.skip(reason="not working")
- def test_supported_memory_backends():
- backends = pa.supported_memory_backends()
-
---- a/pyarrow/tests/parquet/test_basic.py 2023-05-08 09:18:05.307333210 +0200
-+++ b/pyarrow/tests/parquet/test_basic.py 2023-05-08 09:20:16.135429950 +0200
-@@ -349,6 +349,7 @@
- assert result.equals(table)
-
-
-+@pytest.mark.skip(reason="not working")
- @parametrize_legacy_dataset
- def test_byte_stream_split(use_legacy_dataset):
- # This is only a smoke test.
-@@ -510,6 +511,7 @@
- use_legacy_dataset=use_legacy_dataset)
-
-
-+@pytest.mark.skip(reason="not working")
- @parametrize_legacy_dataset
- def test_compression_level(use_legacy_dataset):
- arr = pa.array(list(map(int, range(1000))))
-@@ -660,6 +662,7 @@
- use_legacy_dataset=use_legacy_dataset)
-
-
-+@pytest.mark.skip(reason="not working")
- @pytest.mark.pandas
- @parametrize_legacy_dataset
- def test_zlib_compression_bug(use_legacy_dataset):
-@@ -760,6 +763,7 @@
- assert buf.to_pybytes() == buf.size * b"\0"
-
-
-+@pytest.mark.skip(reason="not working")
- def test_parquet_compression_roundtrip(tempdir):
- # ARROW-10480: ensure even with nonstandard Parquet file naming
- # conventions, writing and then reading a file works. In