Conversation
| return NULL; | ||
| } | ||
| #if defined(__APPLE__) | ||
| shared_mem = 0; |
This comment was marked as off-topic.
This comment was marked as off-topic.
Sorry, something went wrong.
This comment was marked as off-topic.
This comment was marked as off-topic.
Sorry, something went wrong.
This comment was marked as off-topic.
This comment was marked as off-topic.
Sorry, something went wrong.
| const char *filename; | ||
| Py_ssize_t size = 0; | ||
| int shared = 0, shared_mem = 0; | ||
| static char *kwlist[] = {"filename", "shared", "size", "shared_mem", NULL}; |
This comment was marked as off-topic.
This comment was marked as off-topic.
Sorry, something went wrong.
This comment was marked as off-topic.
This comment was marked as off-topic.
Sorry, something went wrong.
This comment was marked as off-topic.
This comment was marked as off-topic.
Sorry, something went wrong.
This comment was marked as off-topic.
This comment was marked as off-topic.
Sorry, something went wrong.
| Py_ssize_t size = 0; | ||
| int shared = 0, shared_mem = 0; | ||
| static char *kwlist[] = {"filename", "shared", "size", "shared_mem", NULL}; | ||
| if (!PyArg_ParseTupleAndKeywords(args, keywds, "s|pnp", kwlist, |
This comment was marked as off-topic.
This comment was marked as off-topic.
Sorry, something went wrong.
test/test_torch.py
Outdated
| def check(filename, share_mem): | ||
| s1 = torch.FloatStorage.from_file(filename, True, size, share_mem) | ||
| t1 = torch.FloatTensor(s1).copy_(torch.randn(size)) | ||
| t1_copy = t1.new().resize_as_(t1).copy_(t1) |
This comment was marked as off-topic.
This comment was marked as off-topic.
Sorry, something went wrong.
test/test_torch.py
Outdated
| s2 = torch.FloatStorage.from_file(filename, True, size) | ||
| t2 = torch.FloatTensor(s2) | ||
| t2_copy = t2.new().resize_as_(t2).copy_(t2) | ||
| self.assertEqual(t1_copy, t2_copy, 0) |
This comment was marked as off-topic.
This comment was marked as off-topic.
Sorry, something went wrong.
| self.assertEqual(t1, t2, 0) | ||
|
|
||
| if sys.startswith('win'): | ||
| os.remove(filename) |
This comment was marked as off-topic.
This comment was marked as off-topic.
Sorry, something went wrong.
test/test_torch.py
Outdated
| def test_from_file(self): | ||
| size = 10000 | ||
| filename = 'testPytorchStorageFromFile' | ||
| sys = platform.system() |
This comment was marked as off-topic.
This comment was marked as off-topic.
Sorry, something went wrong.
test/test_torch.py
Outdated
|
|
||
| # check shm_open + open | ||
| if sys != 'Darwin': # OS X has no /dev/shm | ||
| check(filename, True) |
This comment was marked as off-topic.
This comment was marked as off-topic.
Sorry, something went wrong.
test/test_torch.py
Outdated
| t1 = torch.FloatTensor(s1).copy_(torch.randn(size)) | ||
| t1_copy = t1.new().resize_as_(t1).copy_(t1) | ||
| if share_mem and sys == 'Linux': | ||
| filename = os.path.join('/dev/shm', filename) |
This comment was marked as off-topic.
This comment was marked as off-topic.
Sorry, something went wrong.
This comment was marked as off-topic.
This comment was marked as off-topic.
Sorry, something went wrong.
This comment was marked as off-topic.
This comment was marked as off-topic.
Sorry, something went wrong.
This comment was marked as off-topic.
This comment was marked as off-topic.
Sorry, something went wrong.
|
Actually now I think it should use |
|
So, I'm removing |
|
Sounds good |
|
where should be Storage docs stored? In the _torch_docs.py, or some new _storage_docs.py? |
|
I think you need to create |
|
@apaszke, I've added docs. ready for review/merge |
|
Thank you! |
…2b732d (pytorch#17264) Summary: Pull Request resolved: pytorch#17264 Previous import was 822d8df0a2a32233c6022f50a158817a0f19bdc7 Included changes: - **[4c091e0](onnx/onnx@4c091e0)**: Support defined ONNX_ML in parent cmake files (pytorch#1821) <Lu Fang> - **[57372f3](onnx/onnx@57372f3)**: Delete OpsetVersionConverter.md which is a duplicate of VersionConverter.md (pytorch#1818) <Prasanth Pulavarthi> - **[ab1c57e](onnx/onnx@ab1c57e)**: [ONNXIFI]Add extension to be implementable (pytorch#1796) <Rui Zhu> - **[b92eee8](onnx/onnx@b92eee8)**: Revert "Implement Op Annotation's for ONNX (pytorch#1648)" (pytorch#1812) <Ke Zhang> - **[61f1e9e](onnx/onnx@61f1e9e)**: Enable ONNX_ML by default (pytorch#1810) <Shinichiro Hamaji> - **[4f064a1](onnx/onnx@4f064a1)**: fix Greater and Less doc (pytorch#1811) <Guoliang Hua> - **[0628582](onnx/onnx@0628582)**: Implement Op Annotation's for ONNX (pytorch#1648) <Armen> - **[ad9d2f7](onnx/onnx@ad9d2f7)**: Versioning doc update for Opset 9 (pytorch#1805) <Vinitra Swamy> - **[e71e3be](onnx/onnx@e71e3be)**: add dilation case for ConvTranspose op (pytorch#1797) <Randy> Differential Revision: D14135024 fbshipit-source-id: 9ee7d39a5efea5d9b3c12ac3e6acc32ae83c1d0e
…2b732d (pytorch#17264) Summary: Pull Request resolved: pytorch#17264 Previous import was 822d8df0a2a32233c6022f50a158817a0f19bdc7 Included changes: - **[4c091e0](onnx/onnx@4c091e0)**: Support defined ONNX_ML in parent cmake files (pytorch#1821) <Lu Fang> - **[57372f3](onnx/onnx@57372f3)**: Delete OpsetVersionConverter.md which is a duplicate of VersionConverter.md (pytorch#1818) <Prasanth Pulavarthi> - **[ab1c57e](onnx/onnx@ab1c57e)**: [ONNXIFI]Add extension to be implementable (pytorch#1796) <Rui Zhu> - **[b92eee8](onnx/onnx@b92eee8)**: Revert "Implement Op Annotation's for ONNX (pytorch#1648)" (pytorch#1812) <Ke Zhang> - **[61f1e9e](onnx/onnx@61f1e9e)**: Enable ONNX_ML by default (pytorch#1810) <Shinichiro Hamaji> - **[4f064a1](onnx/onnx@4f064a1)**: fix Greater and Less doc (pytorch#1811) <Guoliang Hua> - **[0628582](onnx/onnx@0628582)**: Implement Op Annotation's for ONNX (pytorch#1648) <Armen> - **[ad9d2f7](onnx/onnx@ad9d2f7)**: Versioning doc update for Opset 9 (pytorch#1805) <Vinitra Swamy> - **[e71e3be](onnx/onnx@e71e3be)**: add dilation case for ConvTranspose op (pytorch#1797) <Randy> Reviewed By: yinghai Differential Revision: D14135024 fbshipit-source-id: 1e4f9dda89abf48994798d080dd5d58207a6e4b6
…pytorch#1821) remove `xfail` from `batch_norm_with_update` op in `test_grad` and `test_vmap_autograd_grad` these tests are passed now Fixes https://ontrack-internal.amd.com/browse/SWDEV-472564 cherry-picked from rocm6.3_internal_testing PR ROCm#1776
…pytorch#1821) remove `xfail` from `batch_norm_with_update` op in `test_grad` and `test_vmap_autograd_grad` these tests are passed now Fixes https://ontrack-internal.amd.com/browse/SWDEV-472564 cherry-picked from rocm6.3_internal_testing PR ROCm#1776 (cherry picked from commit 99b0758)
Resolves #884