mirror of
https://github.com/PaddlePaddle/FastDeploy.git
synced 2025-10-10 11:00:19 +08:00
@@ -638,6 +638,7 @@ if(BUILD_FASTDEPLOY_PYTHON)
|
|||||||
${PYTHON_INCLUDE_DIR})
|
${PYTHON_INCLUDE_DIR})
|
||||||
|
|
||||||
target_include_directories(${PY_LIBRARY_NAME} PUBLIC ${PROJECT_SOURCE_DIR}/third_party/pybind11/include)
|
target_include_directories(${PY_LIBRARY_NAME} PUBLIC ${PROJECT_SOURCE_DIR}/third_party/pybind11/include)
|
||||||
|
target_include_directories(${PY_LIBRARY_NAME} PUBLIC ${PROJECT_SOURCE_DIR}/third_party/dlpack/include)
|
||||||
|
|
||||||
if(APPLE)
|
if(APPLE)
|
||||||
set_target_properties(${PY_LIBRARY_NAME}
|
set_target_properties(${PY_LIBRARY_NAME}
|
||||||
|
39
third_party/dlpack/.github/workflows/deploy.yml
vendored
Normal file
39
third_party/dlpack/.github/workflows/deploy.yml
vendored
Normal file
@@ -0,0 +1,39 @@
|
|||||||
|
name: CI
|
||||||
|
|
||||||
|
on:
|
||||||
|
push:
|
||||||
|
branches:
|
||||||
|
- main
|
||||||
|
|
||||||
|
jobs:
|
||||||
|
test_linux:
|
||||||
|
name: Deploy Docs
|
||||||
|
runs-on: ubuntu-latest
|
||||||
|
|
||||||
|
steps:
|
||||||
|
- uses: actions/checkout@v2
|
||||||
|
with:
|
||||||
|
submodules: recursive
|
||||||
|
|
||||||
|
- name: Configuring Test Environment
|
||||||
|
run: |
|
||||||
|
sudo apt-get update
|
||||||
|
sudo apt-get -y install build-essential doxygen ghp-import
|
||||||
|
python3 -m pip install -U pip wheel
|
||||||
|
|
||||||
|
- name: Installing dependencies
|
||||||
|
run: |
|
||||||
|
python3 -m pip install -r doc_requirements.txt
|
||||||
|
|
||||||
|
- name: Generating Docs
|
||||||
|
run: |
|
||||||
|
make doc
|
||||||
|
|
||||||
|
- name: Deploying on GitHub Pages
|
||||||
|
run: |
|
||||||
|
touch docs/build/.nojekyll
|
||||||
|
git remote set-url origin https://x-access-token:${{ secrets.GITHUB_TOKEN }}@github.com/$GITHUB_REPOSITORY
|
||||||
|
git config --global user.email "dlpack-gh-actions-bot@nomail"
|
||||||
|
git config --global user.name "dlpack-gh-actions-bot"
|
||||||
|
ghp-import -m "Generate DLPack website" -b gh-pages docs/build
|
||||||
|
git push origin gh-pages -f
|
49
third_party/dlpack/.github/workflows/docs.yaml
vendored
Normal file
49
third_party/dlpack/.github/workflows/docs.yaml
vendored
Normal file
@@ -0,0 +1,49 @@
|
|||||||
|
name: Build Doc
|
||||||
|
|
||||||
|
on:
|
||||||
|
push:
|
||||||
|
branches:
|
||||||
|
- main
|
||||||
|
pull_request:
|
||||||
|
branches:
|
||||||
|
- main
|
||||||
|
|
||||||
|
jobs:
|
||||||
|
test_linux:
|
||||||
|
name: Linux
|
||||||
|
runs-on: ubuntu-latest
|
||||||
|
|
||||||
|
steps:
|
||||||
|
- uses: actions/checkout@v2
|
||||||
|
with:
|
||||||
|
submodules: recursive
|
||||||
|
|
||||||
|
- name: Configuring Test Environment
|
||||||
|
run: |
|
||||||
|
sudo apt-get update
|
||||||
|
sudo apt-get -y install build-essential doxygen
|
||||||
|
python3 -m pip install -U pip wheel
|
||||||
|
python3 -m pip install cmake ninja
|
||||||
|
|
||||||
|
python3 --version
|
||||||
|
python3 -m pip --version
|
||||||
|
doxygen --version
|
||||||
|
make --version
|
||||||
|
cmake --version
|
||||||
|
ninja --version
|
||||||
|
|
||||||
|
- name: Installing dependencies
|
||||||
|
run: |
|
||||||
|
python3 -m pip install -r doc_requirements.txt
|
||||||
|
|
||||||
|
- name: Testing CMakeLists.txt
|
||||||
|
run: |
|
||||||
|
mkdir build
|
||||||
|
cd build
|
||||||
|
cmake .. -G Ninja -DCMAKE_INSTALL_PREFIX=./install -DBUILD_DOCS=ON
|
||||||
|
ninja
|
||||||
|
ninja install
|
||||||
|
|
||||||
|
- name: Testing Makefile
|
||||||
|
run: |
|
||||||
|
make doc
|
35
third_party/dlpack/.github/workflows/main.yaml
vendored
Normal file
35
third_party/dlpack/.github/workflows/main.yaml
vendored
Normal file
@@ -0,0 +1,35 @@
|
|||||||
|
name: CI
|
||||||
|
|
||||||
|
on:
|
||||||
|
push:
|
||||||
|
branches:
|
||||||
|
- main
|
||||||
|
pull_request:
|
||||||
|
branches:
|
||||||
|
- main
|
||||||
|
|
||||||
|
jobs:
|
||||||
|
Build:
|
||||||
|
strategy:
|
||||||
|
matrix:
|
||||||
|
os: [ubuntu-latest, macOS-latest]
|
||||||
|
|
||||||
|
runs-on: ${{ matrix.os }}
|
||||||
|
|
||||||
|
steps:
|
||||||
|
- uses: actions/checkout@v2
|
||||||
|
|
||||||
|
- name: Setup Python
|
||||||
|
run: |
|
||||||
|
python3 -m pip install cpplint
|
||||||
|
- name: Setup@Ubuntu
|
||||||
|
if: startsWith(matrix.os, 'ubuntu')
|
||||||
|
run: |
|
||||||
|
sudo apt-get install -y doxygen wget graphviz unzip
|
||||||
|
- name: Lint
|
||||||
|
if: startsWith(matrix.os, 'ubuntu')
|
||||||
|
run: |
|
||||||
|
./tests/scripts/task_lint.sh
|
||||||
|
- name: Test
|
||||||
|
run: |
|
||||||
|
./tests/scripts/task_build.sh
|
32
third_party/dlpack/.gitignore
vendored
Normal file
32
third_party/dlpack/.gitignore
vendored
Normal file
@@ -0,0 +1,32 @@
|
|||||||
|
# Compiled Object files
|
||||||
|
*.slo
|
||||||
|
*.lo
|
||||||
|
*.o
|
||||||
|
*.obj
|
||||||
|
|
||||||
|
# Precompiled Headers
|
||||||
|
*.gch
|
||||||
|
*.pch
|
||||||
|
|
||||||
|
# Compiled Dynamic libraries
|
||||||
|
*.so
|
||||||
|
*.dylib
|
||||||
|
*.dll
|
||||||
|
|
||||||
|
# Fortran module files
|
||||||
|
*.mod
|
||||||
|
*.smod
|
||||||
|
|
||||||
|
# Compiled Static libraries
|
||||||
|
*.lai
|
||||||
|
*.la
|
||||||
|
*.a
|
||||||
|
*.lib
|
||||||
|
|
||||||
|
# Executables
|
||||||
|
*.exe
|
||||||
|
*.out
|
||||||
|
*.app
|
||||||
|
*~
|
||||||
|
build
|
||||||
|
bin
|
127
third_party/dlpack/CMakeLists.txt
vendored
Normal file
127
third_party/dlpack/CMakeLists.txt
vendored
Normal file
@@ -0,0 +1,127 @@
|
|||||||
|
###
|
||||||
|
# Set minimum version of CMake. Since command 'project' use
|
||||||
|
# VERSION sub-option we need at least 3.0.
|
||||||
|
# Note: If you use 2.6 or 2.4, God kills a kitten. Seriously.
|
||||||
|
cmake_minimum_required(VERSION 3.2 FATAL_ERROR)
|
||||||
|
|
||||||
|
####
|
||||||
|
# Set variables:
|
||||||
|
# * PROJECT_NAME
|
||||||
|
# * PROJECT_VERSION
|
||||||
|
project(dlpack VERSION 0.6 LANGUAGES C CXX)
|
||||||
|
|
||||||
|
#####
|
||||||
|
# Change the default build type from Debug to Release, while still
|
||||||
|
# supporting overriding the build type.
|
||||||
|
#
|
||||||
|
# The CACHE STRING logic here and elsewhere is needed to force CMake
|
||||||
|
# to pay attention to the value of these variables.
|
||||||
|
if(NOT CMAKE_BUILD_TYPE)
|
||||||
|
message(STATUS "No build type specified; defaulting to CMAKE_BUILD_TYPE=Release.")
|
||||||
|
set(CMAKE_BUILD_TYPE Release CACHE STRING
|
||||||
|
"Choose the type of build, options are: None Debug Release RelWithDebInfo MinSizeRel."
|
||||||
|
FORCE)
|
||||||
|
else(NOT CMAKE_BUILD_TYPE)
|
||||||
|
if(CMAKE_BUILD_TYPE STREQUAL "Debug")
|
||||||
|
message("==========================================================================================")
|
||||||
|
message(STATUS "Build type: Debug. Performance will be terrible!")
|
||||||
|
message(STATUS "Add -DCMAKE_BUILD_TYPE=Release to the CMake command line to get an optimized build.")
|
||||||
|
message("==========================================================================================")
|
||||||
|
endif(CMAKE_BUILD_TYPE STREQUAL "Debug")
|
||||||
|
endif(NOT CMAKE_BUILD_TYPE)
|
||||||
|
|
||||||
|
####
|
||||||
|
# Setup the compiler options
|
||||||
|
|
||||||
|
# set c++ standard to c++11.
|
||||||
|
# Note: not working on CMake 2.8. We assume that user has
|
||||||
|
# a compiler with C++11 support.
|
||||||
|
|
||||||
|
set(CMAKE_CXX_STANDARD 11)
|
||||||
|
set(CMAKE_CXX_STANDARD_REQUIRED ON)
|
||||||
|
message(STATUS "C++11 support has been enabled by default.")
|
||||||
|
|
||||||
|
option(BUILD_DOCS "Set to ON to build documentation" OFF)
|
||||||
|
option(BUILD_MOCK "Build mock executable" ON)
|
||||||
|
|
||||||
|
if(BUILD_DOCS)
|
||||||
|
add_subdirectory(docs)
|
||||||
|
endif(BUILD_DOCS)
|
||||||
|
|
||||||
|
set(CMAKE_RUNTIME_OUTPUT_DIRECTORY ${PROJECT_SOURCE_DIR}/bin)
|
||||||
|
|
||||||
|
if(BUILD_MOCK)
|
||||||
|
set(DLPACK_MOCK_SRCS ${CMAKE_CURRENT_SOURCE_DIR}/contrib/mock_main.cc
|
||||||
|
${CMAKE_CURRENT_SOURCE_DIR}/contrib/mock_c.c)
|
||||||
|
add_executable(mock ${DLPACK_MOCK_SRCS})
|
||||||
|
endif()
|
||||||
|
|
||||||
|
add_library(dlpack INTERFACE)
|
||||||
|
add_library(${PROJECT_NAME}::dlpack ALIAS dlpack)
|
||||||
|
|
||||||
|
target_include_directories(
|
||||||
|
dlpack
|
||||||
|
INTERFACE
|
||||||
|
$<BUILD_INTERFACE:${CMAKE_CURRENT_SOURCE_DIR}/include>
|
||||||
|
$<BUILD_INTERFACE:${CMAKE_CURRENT_SOURCE_DIR}/contrib>
|
||||||
|
)
|
||||||
|
|
||||||
|
if(BUILD_MOCK)
|
||||||
|
target_link_libraries(mock PRIVATE dlpack)
|
||||||
|
endif()
|
||||||
|
|
||||||
|
# Installation (https://github.com/forexample/package-example) {
|
||||||
|
|
||||||
|
# Introduce variables:
|
||||||
|
# * CMAKE_INSTALL_LIBDIR
|
||||||
|
# * CMAKE_INSTALL_BINDIR
|
||||||
|
# * CMAKE_INSTALL_INCLUDEDIR
|
||||||
|
include(GNUInstallDirs)
|
||||||
|
|
||||||
|
set(config_install_dir "${CMAKE_INSTALL_LIBDIR}/cmake/${PROJECT_NAME}")
|
||||||
|
set(generated_dir "${CMAKE_CURRENT_BINARY_DIR}/generated")
|
||||||
|
set(version_config "${generated_dir}/${PROJECT_NAME}ConfigVersion.cmake")
|
||||||
|
set(project_config "${generated_dir}/${PROJECT_NAME}Config.cmake")
|
||||||
|
set(TARGETS_EXPORT_NAME "${PROJECT_NAME}Targets")
|
||||||
|
set(namespace "${PROJECT_NAME}::")
|
||||||
|
|
||||||
|
include(CMakePackageConfigHelpers)
|
||||||
|
|
||||||
|
# Use:
|
||||||
|
# * PROJECT_VERSION
|
||||||
|
write_basic_package_version_file(
|
||||||
|
"${version_config}" COMPATIBILITY SameMajorVersion
|
||||||
|
)
|
||||||
|
|
||||||
|
# Use:
|
||||||
|
# * TARGETS_EXPORT_NAME
|
||||||
|
# * PROJECT_NAME
|
||||||
|
configure_package_config_file(
|
||||||
|
"cmake/template/Config.cmake.in"
|
||||||
|
"${project_config}"
|
||||||
|
INSTALL_DESTINATION "${config_install_dir}"
|
||||||
|
)
|
||||||
|
|
||||||
|
install(
|
||||||
|
TARGETS dlpack
|
||||||
|
EXPORT "${TARGETS_EXPORT_NAME}"
|
||||||
|
INCLUDES DESTINATION "${CMAKE_INSTALL_INCLUDEDIR}"
|
||||||
|
)
|
||||||
|
|
||||||
|
install(
|
||||||
|
DIRECTORY include/
|
||||||
|
DESTINATION "${CMAKE_INSTALL_INCLUDEDIR}"
|
||||||
|
)
|
||||||
|
|
||||||
|
install(
|
||||||
|
FILES "${project_config}" "${version_config}"
|
||||||
|
DESTINATION "${config_install_dir}"
|
||||||
|
)
|
||||||
|
|
||||||
|
install(
|
||||||
|
EXPORT "${TARGETS_EXPORT_NAME}"
|
||||||
|
NAMESPACE "${namespace}"
|
||||||
|
DESTINATION "${config_install_dir}"
|
||||||
|
)
|
||||||
|
|
||||||
|
# }
|
201
third_party/dlpack/LICENSE
vendored
Normal file
201
third_party/dlpack/LICENSE
vendored
Normal file
@@ -0,0 +1,201 @@
|
|||||||
|
Apache License
|
||||||
|
Version 2.0, January 2004
|
||||||
|
http://www.apache.org/licenses/
|
||||||
|
|
||||||
|
TERMS AND CONDITIONS FOR USE, REPRODUCTION, AND DISTRIBUTION
|
||||||
|
|
||||||
|
1. Definitions.
|
||||||
|
|
||||||
|
"License" shall mean the terms and conditions for use, reproduction,
|
||||||
|
and distribution as defined by Sections 1 through 9 of this document.
|
||||||
|
|
||||||
|
"Licensor" shall mean the copyright owner or entity authorized by
|
||||||
|
the copyright owner that is granting the License.
|
||||||
|
|
||||||
|
"Legal Entity" shall mean the union of the acting entity and all
|
||||||
|
other entities that control, are controlled by, or are under common
|
||||||
|
control with that entity. For the purposes of this definition,
|
||||||
|
"control" means (i) the power, direct or indirect, to cause the
|
||||||
|
direction or management of such entity, whether by contract or
|
||||||
|
otherwise, or (ii) ownership of fifty percent (50%) or more of the
|
||||||
|
outstanding shares, or (iii) beneficial ownership of such entity.
|
||||||
|
|
||||||
|
"You" (or "Your") shall mean an individual or Legal Entity
|
||||||
|
exercising permissions granted by this License.
|
||||||
|
|
||||||
|
"Source" form shall mean the preferred form for making modifications,
|
||||||
|
including but not limited to software source code, documentation
|
||||||
|
source, and configuration files.
|
||||||
|
|
||||||
|
"Object" form shall mean any form resulting from mechanical
|
||||||
|
transformation or translation of a Source form, including but
|
||||||
|
not limited to compiled object code, generated documentation,
|
||||||
|
and conversions to other media types.
|
||||||
|
|
||||||
|
"Work" shall mean the work of authorship, whether in Source or
|
||||||
|
Object form, made available under the License, as indicated by a
|
||||||
|
copyright notice that is included in or attached to the work
|
||||||
|
(an example is provided in the Appendix below).
|
||||||
|
|
||||||
|
"Derivative Works" shall mean any work, whether in Source or Object
|
||||||
|
form, that is based on (or derived from) the Work and for which the
|
||||||
|
editorial revisions, annotations, elaborations, or other modifications
|
||||||
|
represent, as a whole, an original work of authorship. For the purposes
|
||||||
|
of this License, Derivative Works shall not include works that remain
|
||||||
|
separable from, or merely link (or bind by name) to the interfaces of,
|
||||||
|
the Work and Derivative Works thereof.
|
||||||
|
|
||||||
|
"Contribution" shall mean any work of authorship, including
|
||||||
|
the original version of the Work and any modifications or additions
|
||||||
|
to that Work or Derivative Works thereof, that is intentionally
|
||||||
|
submitted to Licensor for inclusion in the Work by the copyright owner
|
||||||
|
or by an individual or Legal Entity authorized to submit on behalf of
|
||||||
|
the copyright owner. For the purposes of this definition, "submitted"
|
||||||
|
means any form of electronic, verbal, or written communication sent
|
||||||
|
to the Licensor or its representatives, including but not limited to
|
||||||
|
communication on electronic mailing lists, source code control systems,
|
||||||
|
and issue tracking systems that are managed by, or on behalf of, the
|
||||||
|
Licensor for the purpose of discussing and improving the Work, but
|
||||||
|
excluding communication that is conspicuously marked or otherwise
|
||||||
|
designated in writing by the copyright owner as "Not a Contribution."
|
||||||
|
|
||||||
|
"Contributor" shall mean Licensor and any individual or Legal Entity
|
||||||
|
on behalf of whom a Contribution has been received by Licensor and
|
||||||
|
subsequently incorporated within the Work.
|
||||||
|
|
||||||
|
2. Grant of Copyright License. Subject to the terms and conditions of
|
||||||
|
this License, each Contributor hereby grants to You a perpetual,
|
||||||
|
worldwide, non-exclusive, no-charge, royalty-free, irrevocable
|
||||||
|
copyright license to reproduce, prepare Derivative Works of,
|
||||||
|
publicly display, publicly perform, sublicense, and distribute the
|
||||||
|
Work and such Derivative Works in Source or Object form.
|
||||||
|
|
||||||
|
3. Grant of Patent License. Subject to the terms and conditions of
|
||||||
|
this License, each Contributor hereby grants to You a perpetual,
|
||||||
|
worldwide, non-exclusive, no-charge, royalty-free, irrevocable
|
||||||
|
(except as stated in this section) patent license to make, have made,
|
||||||
|
use, offer to sell, sell, import, and otherwise transfer the Work,
|
||||||
|
where such license applies only to those patent claims licensable
|
||||||
|
by such Contributor that are necessarily infringed by their
|
||||||
|
Contribution(s) alone or by combination of their Contribution(s)
|
||||||
|
with the Work to which such Contribution(s) was submitted. If You
|
||||||
|
institute patent litigation against any entity (including a
|
||||||
|
cross-claim or counterclaim in a lawsuit) alleging that the Work
|
||||||
|
or a Contribution incorporated within the Work constitutes direct
|
||||||
|
or contributory patent infringement, then any patent licenses
|
||||||
|
granted to You under this License for that Work shall terminate
|
||||||
|
as of the date such litigation is filed.
|
||||||
|
|
||||||
|
4. Redistribution. You may reproduce and distribute copies of the
|
||||||
|
Work or Derivative Works thereof in any medium, with or without
|
||||||
|
modifications, and in Source or Object form, provided that You
|
||||||
|
meet the following conditions:
|
||||||
|
|
||||||
|
(a) You must give any other recipients of the Work or
|
||||||
|
Derivative Works a copy of this License; and
|
||||||
|
|
||||||
|
(b) You must cause any modified files to carry prominent notices
|
||||||
|
stating that You changed the files; and
|
||||||
|
|
||||||
|
(c) You must retain, in the Source form of any Derivative Works
|
||||||
|
that You distribute, all copyright, patent, trademark, and
|
||||||
|
attribution notices from the Source form of the Work,
|
||||||
|
excluding those notices that do not pertain to any part of
|
||||||
|
the Derivative Works; and
|
||||||
|
|
||||||
|
(d) If the Work includes a "NOTICE" text file as part of its
|
||||||
|
distribution, then any Derivative Works that You distribute must
|
||||||
|
include a readable copy of the attribution notices contained
|
||||||
|
within such NOTICE file, excluding those notices that do not
|
||||||
|
pertain to any part of the Derivative Works, in at least one
|
||||||
|
of the following places: within a NOTICE text file distributed
|
||||||
|
as part of the Derivative Works; within the Source form or
|
||||||
|
documentation, if provided along with the Derivative Works; or,
|
||||||
|
within a display generated by the Derivative Works, if and
|
||||||
|
wherever such third-party notices normally appear. The contents
|
||||||
|
of the NOTICE file are for informational purposes only and
|
||||||
|
do not modify the License. You may add Your own attribution
|
||||||
|
notices within Derivative Works that You distribute, alongside
|
||||||
|
or as an addendum to the NOTICE text from the Work, provided
|
||||||
|
that such additional attribution notices cannot be construed
|
||||||
|
as modifying the License.
|
||||||
|
|
||||||
|
You may add Your own copyright statement to Your modifications and
|
||||||
|
may provide additional or different license terms and conditions
|
||||||
|
for use, reproduction, or distribution of Your modifications, or
|
||||||
|
for any such Derivative Works as a whole, provided Your use,
|
||||||
|
reproduction, and distribution of the Work otherwise complies with
|
||||||
|
the conditions stated in this License.
|
||||||
|
|
||||||
|
5. Submission of Contributions. Unless You explicitly state otherwise,
|
||||||
|
any Contribution intentionally submitted for inclusion in the Work
|
||||||
|
by You to the Licensor shall be under the terms and conditions of
|
||||||
|
this License, without any additional terms or conditions.
|
||||||
|
Notwithstanding the above, nothing herein shall supersede or modify
|
||||||
|
the terms of any separate license agreement you may have executed
|
||||||
|
with Licensor regarding such Contributions.
|
||||||
|
|
||||||
|
6. Trademarks. This License does not grant permission to use the trade
|
||||||
|
names, trademarks, service marks, or product names of the Licensor,
|
||||||
|
except as required for reasonable and customary use in describing the
|
||||||
|
origin of the Work and reproducing the content of the NOTICE file.
|
||||||
|
|
||||||
|
7. Disclaimer of Warranty. Unless required by applicable law or
|
||||||
|
agreed to in writing, Licensor provides the Work (and each
|
||||||
|
Contributor provides its Contributions) on an "AS IS" BASIS,
|
||||||
|
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or
|
||||||
|
implied, including, without limitation, any warranties or conditions
|
||||||
|
of TITLE, NON-INFRINGEMENT, MERCHANTABILITY, or FITNESS FOR A
|
||||||
|
PARTICULAR PURPOSE. You are solely responsible for determining the
|
||||||
|
appropriateness of using or redistributing the Work and assume any
|
||||||
|
risks associated with Your exercise of permissions under this License.
|
||||||
|
|
||||||
|
8. Limitation of Liability. In no event and under no legal theory,
|
||||||
|
whether in tort (including negligence), contract, or otherwise,
|
||||||
|
unless required by applicable law (such as deliberate and grossly
|
||||||
|
negligent acts) or agreed to in writing, shall any Contributor be
|
||||||
|
liable to You for damages, including any direct, indirect, special,
|
||||||
|
incidental, or consequential damages of any character arising as a
|
||||||
|
result of this License or out of the use or inability to use the
|
||||||
|
Work (including but not limited to damages for loss of goodwill,
|
||||||
|
work stoppage, computer failure or malfunction, or any and all
|
||||||
|
other commercial damages or losses), even if such Contributor
|
||||||
|
has been advised of the possibility of such damages.
|
||||||
|
|
||||||
|
9. Accepting Warranty or Additional Liability. While redistributing
|
||||||
|
the Work or Derivative Works thereof, You may choose to offer,
|
||||||
|
and charge a fee for, acceptance of support, warranty, indemnity,
|
||||||
|
or other liability obligations and/or rights consistent with this
|
||||||
|
License. However, in accepting such obligations, You may act only
|
||||||
|
on Your own behalf and on Your sole responsibility, not on behalf
|
||||||
|
of any other Contributor, and only if You agree to indemnify,
|
||||||
|
defend, and hold each Contributor harmless for any liability
|
||||||
|
incurred by, or claims asserted against, such Contributor by reason
|
||||||
|
of your accepting any such warranty or additional liability.
|
||||||
|
|
||||||
|
END OF TERMS AND CONDITIONS
|
||||||
|
|
||||||
|
APPENDIX: How to apply the Apache License to your work.
|
||||||
|
|
||||||
|
To apply the Apache License to your work, attach the following
|
||||||
|
boilerplate notice, with the fields enclosed by brackets "{}"
|
||||||
|
replaced with your own identifying information. (Don't include
|
||||||
|
the brackets!) The text should be enclosed in the appropriate
|
||||||
|
comment syntax for the file format. We also recommend that a
|
||||||
|
file or class name and description of purpose be included on the
|
||||||
|
same "printed page" as the copyright notice for easier
|
||||||
|
identification within third-party archives.
|
||||||
|
|
||||||
|
Copyright 2017 by Contributors
|
||||||
|
|
||||||
|
Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
|
you may not use this file except in compliance with the License.
|
||||||
|
You may obtain a copy of the License at
|
||||||
|
|
||||||
|
http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
|
||||||
|
Unless required by applicable law or agreed to in writing, software
|
||||||
|
distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
See the License for the specific language governing permissions and
|
||||||
|
limitations under the License.
|
39
third_party/dlpack/Makefile
vendored
Normal file
39
third_party/dlpack/Makefile
vendored
Normal file
@@ -0,0 +1,39 @@
|
|||||||
|
.PHONY: clean all test doc lint show_docs
|
||||||
|
|
||||||
|
all: bin/mock
|
||||||
|
|
||||||
|
LDFLAGS =
|
||||||
|
CFLAGS = -Wall -O3 -Iinclude -Icontrib
|
||||||
|
CXXFLAGS = -std=c++11 $(CFLAGS)
|
||||||
|
|
||||||
|
SRC = $(wildcard contrib/*.cc contrib/*.c)
|
||||||
|
ALL_CXX_OBJ = $(patsubst contrib/%.cc, build/%.o, $(SRC))
|
||||||
|
ALL_C_OBJ = $(patsubst contrib/%.c, build/%.o, $(SRC))
|
||||||
|
ALL_OBJ = $(ALL_CC_OBJ) $(ALL_CXX_OBJ)
|
||||||
|
|
||||||
|
doc:
|
||||||
|
doxygen docs/Doxyfile
|
||||||
|
$(MAKE) -C docs html
|
||||||
|
|
||||||
|
show_docs:
|
||||||
|
@python3 -m http.server --directory docs/build/latest
|
||||||
|
|
||||||
|
lint:
|
||||||
|
./tests/scripts/task_lint.sh
|
||||||
|
|
||||||
|
build/%.o: contrib/%.cc
|
||||||
|
@mkdir -p $(@D)
|
||||||
|
$(CXX) $(CXXFLAGS) -MM -MT build/$*.o $< >build/$*.d
|
||||||
|
$(CXX) -c $(CXXFLAGS) -c $< -o $@
|
||||||
|
|
||||||
|
build/%.o: contrib/%.c
|
||||||
|
@mkdir -p $(@D)
|
||||||
|
$(CC) $(CFLAGS) -MM -MT build/$*.o $< >build/$*.d
|
||||||
|
$(CC) -c $(CFLAGS) -c $< -o $@
|
||||||
|
|
||||||
|
bin/mock: $(ALL_OBJ)
|
||||||
|
@mkdir -p $(@D)
|
||||||
|
$(CXX) $(CXXFLAGS) -o $@ $(filter %.o %.a, $^) $(LDFLAGS)
|
||||||
|
|
||||||
|
clean:
|
||||||
|
$(RM) -rf build */*/*/*~ */*.o */*/*.o */*/*/*.o */*.d */*/*.d */*/*/*.d docs/build docs/doxygen
|
53
third_party/dlpack/NEWS.md
vendored
Normal file
53
third_party/dlpack/NEWS.md
vendored
Normal file
@@ -0,0 +1,53 @@
|
|||||||
|
DLPack Change Log
|
||||||
|
=================
|
||||||
|
|
||||||
|
This file records the changes in DLPack in reverse chronological order.
|
||||||
|
|
||||||
|
## v0.7
|
||||||
|
|
||||||
|
- Add kDLHexagon
|
||||||
|
- Add kDLOneAPI
|
||||||
|
- Add DLPACK_VERSION and DLPACK_ABI_VERSION
|
||||||
|
|
||||||
|
|
||||||
|
## v0.6
|
||||||
|
|
||||||
|
- Add kDLROCMHost
|
||||||
|
- Add kDLCUDAManaged
|
||||||
|
|
||||||
|
|
||||||
|
## v0.5
|
||||||
|
|
||||||
|
Rename enum values
|
||||||
|
- kDLGPU -> kDLCUDA
|
||||||
|
- kDLCPUPinned -> kDLCUDAHost
|
||||||
|
The ABI is backward compatible, as it is only change of constant name,
|
||||||
|
exchange can still happen between the new version and old version.
|
||||||
|
|
||||||
|
|
||||||
|
## v0.4
|
||||||
|
|
||||||
|
- OpaqueHandle type
|
||||||
|
- Complex support
|
||||||
|
- Rename DLContext -> DLDevice
|
||||||
|
- This requires dependent frameworks to upgrade the type name.
|
||||||
|
- The ABI is backward compatible, as it is only change of constant name.
|
||||||
|
|
||||||
|
## v0.3
|
||||||
|
|
||||||
|
- Add bfloat16
|
||||||
|
- Vulkan support
|
||||||
|
|
||||||
|
|
||||||
|
## v0.2
|
||||||
|
- New device types
|
||||||
|
- kDLMetal for Apple Metal device
|
||||||
|
- kDLVPI for verilog simulator memory
|
||||||
|
- kDLROCM for AMD GPUs
|
||||||
|
- Add prefix DL to all enum constant values
|
||||||
|
- This requires dependent frameworks to upgrade their reference to these constant
|
||||||
|
- The ABI is compatible, as it is only change of constant name.
|
||||||
|
- Add DLManagedTensor structure for borrowing tensors
|
||||||
|
|
||||||
|
## v0.1
|
||||||
|
- Finalize DLTensor structure
|
29
third_party/dlpack/README.md
vendored
Normal file
29
third_party/dlpack/README.md
vendored
Normal file
@@ -0,0 +1,29 @@
|
|||||||
|
# DLPack: Open In Memory Tensor Structure
|
||||||
|
|
||||||
|
[](https://github.com/dmlc/dlpack/actions/workflows/main.yaml)
|
||||||
|
|
||||||
|
Documentation: [https://dmlc.github.io/dlpack/latest](https://dmlc.github.io/dlpack/latest)
|
||||||
|
|
||||||
|
DLPack is an open in-memory tensor structure for sharing tensors among frameworks. DLPack enables
|
||||||
|
|
||||||
|
- Easier sharing of operators between deep learning frameworks.
|
||||||
|
- Easier wrapping of vendor level operator implementations, allowing collaboration when introducing new devices/ops.
|
||||||
|
- Quick swapping of backend implementations, like different version of BLAS
|
||||||
|
- For final users, this could bring more operators, and possibility of mixing usage between frameworks.
|
||||||
|
|
||||||
|
We do not intend to implement Tensor and Ops, but instead use this as common bridge
|
||||||
|
to reuse tensor and ops across frameworks.
|
||||||
|
|
||||||
|
## Proposal Procedure
|
||||||
|
RFC proposals are opened as issues. The major release will happen as a vote issue to make
|
||||||
|
sure the participants agree on the changes.
|
||||||
|
|
||||||
|
## Project Structure
|
||||||
|
There are two major components so far
|
||||||
|
- include: stabilized headers
|
||||||
|
- contrib: in progress unstable libraries
|
||||||
|
|
||||||
|
## People
|
||||||
|
Here are list of people who have been involved in DLPack RFC design proposals:
|
||||||
|
|
||||||
|
@soumith @piiswrong @Yangqing @naibaf7 @bhack @edgarriba @tqchen @prigoyal @zdevito
|
1
third_party/dlpack/apps/numpy_dlpack/.gitignore
vendored
Normal file
1
third_party/dlpack/apps/numpy_dlpack/.gitignore
vendored
Normal file
@@ -0,0 +1 @@
|
|||||||
|
__pycache__
|
23
third_party/dlpack/apps/numpy_dlpack/README.md
vendored
Normal file
23
third_party/dlpack/apps/numpy_dlpack/README.md
vendored
Normal file
@@ -0,0 +1,23 @@
|
|||||||
|
# Numpy DLPack Array Conversion Example
|
||||||
|
|
||||||
|
This example demonstrates how a underlying array memory can be handed off between
|
||||||
|
two DLPack compatible frameworks without requiring any copies. In this case,
|
||||||
|
we demonstrate how to convert numpy to TVM's NDArray and vice-versa with proper
|
||||||
|
memory handling. We hope that not only is this directly useful for TVM users, but
|
||||||
|
also a solid example for how similar efficient copies can be implemented in other
|
||||||
|
array frameworks.
|
||||||
|
|
||||||
|
## File Breakdown
|
||||||
|
|
||||||
|
[dlpack.py](dlpack/dlpack.py): Contains the definition of common DLPack structures shared between frameworks. Mirrors the official C++ definitions.
|
||||||
|
|
||||||
|
[from_numpy.py](dlpack/from_numpy.py): Demonstrates how to convert a numpy array into a PyCapsule containing a DLPack Tensor.
|
||||||
|
|
||||||
|
[to_numpy.py](dlpack/to_numpy.py): Demonstrates how to take a PyCapsule with a DLPack Tensor and convert it into a numpy array.
|
||||||
|
|
||||||
|
[test.py](dlpack/test.py): Shows how to_numpy and from_numpy can be used to convert tensor formats without copies.
|
||||||
|
|
||||||
|
## Authors
|
||||||
|
[Josh Fromm](https://github.com/jwfromm)
|
||||||
|
|
||||||
|
[Junru Shao](https://github.com/junrushao1994)
|
2
third_party/dlpack/apps/numpy_dlpack/dlpack/__init__.py
vendored
Normal file
2
third_party/dlpack/apps/numpy_dlpack/dlpack/__init__.py
vendored
Normal file
@@ -0,0 +1,2 @@
|
|||||||
|
from .from_numpy import from_numpy
|
||||||
|
from .to_numpy import to_numpy
|
137
third_party/dlpack/apps/numpy_dlpack/dlpack/dlpack.py
vendored
Normal file
137
third_party/dlpack/apps/numpy_dlpack/dlpack/dlpack.py
vendored
Normal file
@@ -0,0 +1,137 @@
|
|||||||
|
import ctypes
|
||||||
|
|
||||||
|
_c_str_dltensor = b"dltensor"
|
||||||
|
|
||||||
|
|
||||||
|
class DLDeviceType(ctypes.c_int):
|
||||||
|
"""The enum that encodes the type of the device where
|
||||||
|
DLTensor memory is allocated.
|
||||||
|
"""
|
||||||
|
kDLCPU = 1
|
||||||
|
kDLCUDA = 2
|
||||||
|
kDLCUDAHost = 3
|
||||||
|
kDLOpenCL = 4
|
||||||
|
kDLVulkan = 7
|
||||||
|
kDLMetal = 8
|
||||||
|
kDLVPI = 9
|
||||||
|
kDLROCM = 10
|
||||||
|
kDLROCMHost = 11
|
||||||
|
kDLCUDAManaged = 13
|
||||||
|
kDLOneAPI = 14
|
||||||
|
|
||||||
|
def __str__(self):
|
||||||
|
return {
|
||||||
|
self.kDLCPU: "CPU",
|
||||||
|
self.kDLCUDA: "CUDA",
|
||||||
|
self.kDLCUDAHost: "CUDAHost",
|
||||||
|
self.kDLOpenCL: "OpenCL",
|
||||||
|
self.kDLVulkan: "Vulkan",
|
||||||
|
self.kDLMetal: "Metal",
|
||||||
|
self.kDLVPI: "VPI",
|
||||||
|
self.kDLROCM: "ROCM",
|
||||||
|
self.kDLROCMHost: "ROMCHost",
|
||||||
|
self.kDLCUDAManaged: "CUDAManaged",
|
||||||
|
self.kDLOneAPI: "oneAPI",
|
||||||
|
}[self.value]
|
||||||
|
|
||||||
|
|
||||||
|
class DLDevice(ctypes.Structure):
|
||||||
|
"""Represents the device where DLTensor memory is allocated.
|
||||||
|
The device is represented by the pair of fields:
|
||||||
|
device_type: DLDeviceType
|
||||||
|
device_id: c_int
|
||||||
|
"""
|
||||||
|
_fields_ = [
|
||||||
|
("device_type", DLDeviceType),
|
||||||
|
("device_id", ctypes.c_int),
|
||||||
|
]
|
||||||
|
|
||||||
|
|
||||||
|
class DLDataTypeCode(ctypes.c_uint8):
|
||||||
|
"""An integer that encodes the category of DLTensor elements' data type."""
|
||||||
|
kDLInt = 0
|
||||||
|
kDLUInt = 1
|
||||||
|
kDLFloat = 2
|
||||||
|
kDLOpaquePointer = 3
|
||||||
|
kDLBfloat = 4
|
||||||
|
kDLComplex = 5
|
||||||
|
|
||||||
|
def __str__(self):
|
||||||
|
return {
|
||||||
|
self.kDLInt: "int",
|
||||||
|
self.kDLUInt: "uint",
|
||||||
|
self.kDLFloat: "float",
|
||||||
|
self.kDLBfloat: "bfloat",
|
||||||
|
self.kDLComplex: "complex",
|
||||||
|
self.kDLOpaquePointer: "void_p"
|
||||||
|
}[self.value]
|
||||||
|
|
||||||
|
|
||||||
|
class DLDataType(ctypes.Structure):
|
||||||
|
"""Descriptor of data type for elements of DLTensor.
|
||||||
|
The data type is described by a triple, `DLDataType.type_code`,
|
||||||
|
`DLDataType.bits`, and `DLDataType.lanes`.
|
||||||
|
|
||||||
|
The element is understood as packed `lanes` repetitions of
|
||||||
|
elements from `type_code` data-category of width `bits`.
|
||||||
|
"""
|
||||||
|
_fields_ = [
|
||||||
|
("type_code", DLDataTypeCode),
|
||||||
|
("bits", ctypes.c_uint8),
|
||||||
|
("lanes", ctypes.c_uint16),
|
||||||
|
]
|
||||||
|
TYPE_MAP = {
|
||||||
|
"bool": (DLDataTypeCode.kDLUInt, 1, 1),
|
||||||
|
"int8": (DLDataTypeCode.kDLInt, 8, 1),
|
||||||
|
"int16": (DLDataTypeCode.kDLInt, 16, 1),
|
||||||
|
"int32": (DLDataTypeCode.kDLInt, 32, 1),
|
||||||
|
"int64": (DLDataTypeCode.kDLInt, 64, 1),
|
||||||
|
"uint8": (DLDataTypeCode.kDLUInt, 8, 1),
|
||||||
|
"uint16": (DLDataTypeCode.kDLUInt, 16, 1),
|
||||||
|
"uint32": (DLDataTypeCode.kDLUInt, 32, 1),
|
||||||
|
"uint64": (DLDataTypeCode.kDLUInt, 64, 1),
|
||||||
|
"float16": (DLDataTypeCode.kDLFloat, 16, 1),
|
||||||
|
"float32": (DLDataTypeCode.kDLFloat, 32, 1),
|
||||||
|
"float64": (DLDataTypeCode.kDLFloat, 64, 1),
|
||||||
|
"complex64": (DLDataTypeCode.kDLComplex, 64, 1),
|
||||||
|
"complex128": (DLDataTypeCode.kDLComplex, 128, 1)
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
class DLTensor(ctypes.Structure):
|
||||||
|
"""Structure describing strided layout of DLTensor.
|
||||||
|
Fields are:
|
||||||
|
data: void pointer
|
||||||
|
device: DLDevice
|
||||||
|
ndim: number of indices needed to reference an
|
||||||
|
element of the tensor
|
||||||
|
dtype: data type descriptor
|
||||||
|
shape: tuple with lengths of the corresponding
|
||||||
|
tensor dimensions
|
||||||
|
strides: tuple of numbers of array elements to
|
||||||
|
step in each dimension when traversing
|
||||||
|
the tensor
|
||||||
|
byte_offset: data + byte_offset gives the address of
|
||||||
|
tensor element with index (0,) * ndim
|
||||||
|
"""
|
||||||
|
_fields_ = [
|
||||||
|
("data", ctypes.c_void_p),
|
||||||
|
("device", DLDevice),
|
||||||
|
("ndim", ctypes.c_int),
|
||||||
|
("dtype", DLDataType),
|
||||||
|
("shape", ctypes.POINTER(ctypes.c_int64)),
|
||||||
|
("strides", ctypes.POINTER(ctypes.c_int64)),
|
||||||
|
("byte_offset", ctypes.c_uint64),
|
||||||
|
]
|
||||||
|
|
||||||
|
|
||||||
|
class DLManagedTensor(ctypes.Structure):
|
||||||
|
"""Structure storing the pointer to the tensor descriptor,
|
||||||
|
deleter callable for the tensor descriptor, and pointer to
|
||||||
|
some additional data. These are stored in fields `dl_tensor`,
|
||||||
|
`deleter`, and `manager_ctx`."""
|
||||||
|
_fields_ = [
|
||||||
|
("dl_tensor", DLTensor),
|
||||||
|
("manager_ctx", ctypes.c_void_p),
|
||||||
|
("deleter", ctypes.CFUNCTYPE(None, ctypes.c_void_p)),
|
||||||
|
]
|
96
third_party/dlpack/apps/numpy_dlpack/dlpack/from_numpy.py
vendored
Normal file
96
third_party/dlpack/apps/numpy_dlpack/dlpack/from_numpy.py
vendored
Normal file
@@ -0,0 +1,96 @@
|
|||||||
|
from typing import Callable
|
||||||
|
import numpy as np
|
||||||
|
import ctypes
|
||||||
|
|
||||||
|
from .dlpack import DLManagedTensor, DLDevice, DLDataType, _c_str_dltensor
|
||||||
|
|
||||||
|
ctypes.pythonapi.PyMem_RawMalloc.restype = ctypes.c_void_p
|
||||||
|
ctypes.pythonapi.PyMem_RawFree.argtypes = [ctypes.c_void_p]
|
||||||
|
|
||||||
|
ctypes.pythonapi.PyCapsule_New.restype = ctypes.py_object
|
||||||
|
ctypes.pythonapi.PyCapsule_New.argtypes = [
|
||||||
|
ctypes.c_void_p, ctypes.c_char_p, ctypes.c_void_p
|
||||||
|
]
|
||||||
|
|
||||||
|
|
||||||
|
class _Holder:
|
||||||
|
"""A wrapper around a numpy array to keep track of references to the underlying memory.
|
||||||
|
|
||||||
|
Parameters
|
||||||
|
----------
|
||||||
|
np_array : np.ndarray
|
||||||
|
The numpy array that will be converted to a DLPack tensor and must be managed.
|
||||||
|
"""
|
||||||
|
|
||||||
|
def __init__(self, np_array: np.ndarray) -> None:
|
||||||
|
self.np_array = np_array
|
||||||
|
self.data = np_array.ctypes.data_as(ctypes.c_void_p)
|
||||||
|
self.shape = np_array.ctypes.shape_as(ctypes.c_int64)
|
||||||
|
self.strides = np_array.ctypes.strides_as(ctypes.c_int64)
|
||||||
|
for i in range(np_array.ndim):
|
||||||
|
self.strides[i] //= np_array.itemsize
|
||||||
|
|
||||||
|
def _as_manager_ctx(self) -> ctypes.c_void_p:
|
||||||
|
py_obj = ctypes.py_object(self)
|
||||||
|
py_obj_ptr = ctypes.pointer(py_obj)
|
||||||
|
ctypes.pythonapi.Py_IncRef(py_obj)
|
||||||
|
ctypes.pythonapi.Py_IncRef(ctypes.py_object(py_obj_ptr))
|
||||||
|
return ctypes.cast(py_obj_ptr, ctypes.c_void_p)
|
||||||
|
|
||||||
|
|
||||||
|
@ctypes.CFUNCTYPE(None, ctypes.c_void_p)
|
||||||
|
def _numpy_array_deleter(handle: ctypes.c_void_p) -> None:
|
||||||
|
"""A function to deallocate the memory of a numpy array."""
|
||||||
|
dl_managed_tensor = DLManagedTensor.from_address(handle)
|
||||||
|
py_obj_ptr = ctypes.cast(dl_managed_tensor.manager_ctx,
|
||||||
|
ctypes.POINTER(ctypes.py_object))
|
||||||
|
py_obj = py_obj_ptr.contents
|
||||||
|
ctypes.pythonapi.Py_DecRef(py_obj)
|
||||||
|
ctypes.pythonapi.Py_DecRef(ctypes.py_object(py_obj_ptr))
|
||||||
|
ctypes.pythonapi.PyMem_RawFree(handle)
|
||||||
|
|
||||||
|
|
||||||
|
@ctypes.CFUNCTYPE(None, ctypes.c_void_p)
|
||||||
|
def _numpy_pycapsule_deleter(handle: ctypes.c_void_p) -> None:
|
||||||
|
"""A function to deallocate a pycapsule that wraps a numpy array."""
|
||||||
|
pycapsule: ctypes.py_object = ctypes.cast(handle, ctypes.py_object)
|
||||||
|
if ctypes.pythonapi.PyCapsule_IsValid(pycapsule, _c_str_dltensor):
|
||||||
|
dl_managed_tensor = ctypes.pythonapi.PyCapsule_GetPointer(
|
||||||
|
pycapsule, _c_str_dltensor)
|
||||||
|
_numpy_array_deleter(dl_managed_tensor)
|
||||||
|
ctypes.pythonapi.PyCapsule_SetDestructor(pycapsule, None)
|
||||||
|
|
||||||
|
|
||||||
|
def from_numpy(np_array: np.ndarray):
|
||||||
|
"""Convert a numpy array to another type of dlpack compatible array.
|
||||||
|
|
||||||
|
Parameters
|
||||||
|
----------
|
||||||
|
np_array : np.ndarray
|
||||||
|
The source numpy array that will be converted.
|
||||||
|
|
||||||
|
Returns
|
||||||
|
-------
|
||||||
|
pycapsule : PyCapsule
|
||||||
|
A pycapsule containing a DLManagedTensor that can be converted
|
||||||
|
to other array formats without copying the underlying memory.
|
||||||
|
"""
|
||||||
|
holder = _Holder(np_array)
|
||||||
|
size = ctypes.c_size_t(ctypes.sizeof(DLManagedTensor))
|
||||||
|
dl_managed_tensor = DLManagedTensor.from_address(
|
||||||
|
ctypes.pythonapi.PyMem_RawMalloc(size))
|
||||||
|
dl_managed_tensor.dl_tensor.data = holder.data
|
||||||
|
dl_managed_tensor.dl_tensor.device = DLDevice(1, 0)
|
||||||
|
dl_managed_tensor.dl_tensor.ndim = np_array.ndim
|
||||||
|
dl_managed_tensor.dl_tensor.dtype = DLDataType.TYPE_MAP[str(
|
||||||
|
np_array.dtype)]
|
||||||
|
dl_managed_tensor.dl_tensor.shape = holder.shape
|
||||||
|
dl_managed_tensor.dl_tensor.strides = holder.strides
|
||||||
|
dl_managed_tensor.dl_tensor.byte_offset = 0
|
||||||
|
dl_managed_tensor.manager_ctx = holder._as_manager_ctx()
|
||||||
|
dl_managed_tensor.deleter = _numpy_array_deleter
|
||||||
|
pycapsule = ctypes.pythonapi.PyCapsule_New(
|
||||||
|
ctypes.byref(dl_managed_tensor),
|
||||||
|
_c_str_dltensor,
|
||||||
|
_numpy_pycapsule_deleter, )
|
||||||
|
return pycapsule
|
83
third_party/dlpack/apps/numpy_dlpack/dlpack/to_numpy.py
vendored
Normal file
83
third_party/dlpack/apps/numpy_dlpack/dlpack/to_numpy.py
vendored
Normal file
@@ -0,0 +1,83 @@
|
|||||||
|
import ctypes
|
||||||
|
import numpy as np
|
||||||
|
from .dlpack import _c_str_dltensor, DLManagedTensor, DLTensor
|
||||||
|
|
||||||
|
ctypes.pythonapi.PyCapsule_IsValid.restype = ctypes.c_int
|
||||||
|
ctypes.pythonapi.PyCapsule_IsValid.argtypes = [
|
||||||
|
ctypes.py_object, ctypes.c_char_p
|
||||||
|
]
|
||||||
|
|
||||||
|
ctypes.pythonapi.PyCapsule_GetPointer.restype = ctypes.c_void_p
|
||||||
|
ctypes.pythonapi.PyCapsule_GetPointer.argtypes = [
|
||||||
|
ctypes.py_object, ctypes.c_char_p
|
||||||
|
]
|
||||||
|
|
||||||
|
|
||||||
|
def _array_interface_from_dl_tensor(dlt):
|
||||||
|
"""Constructs NumPy's array_interface dictionary
|
||||||
|
from `dlpack.DLTensor` descriptor."""
|
||||||
|
assert isinstance(dlt, DLTensor)
|
||||||
|
shape = tuple(dlt.shape[dim] for dim in range(dlt.ndim))
|
||||||
|
itemsize = dlt.dtype.lanes * dlt.dtype.bits // 8
|
||||||
|
if dlt.strides:
|
||||||
|
strides = tuple(dlt.strides[dim] * itemsize for dim in range(dlt.ndim))
|
||||||
|
else:
|
||||||
|
# Array is compact, make it numpy compatible.
|
||||||
|
strides = []
|
||||||
|
for i, s in enumerate(shape):
|
||||||
|
cumulative = 1
|
||||||
|
for e in range(i + 1, dlt.ndim):
|
||||||
|
cumulative *= shape[e]
|
||||||
|
strides.append(cumulative * itemsize)
|
||||||
|
strides = tuple(strides)
|
||||||
|
typestr = "|" + str(dlt.dtype.type_code)[0] + str(itemsize)
|
||||||
|
return dict(
|
||||||
|
version=3,
|
||||||
|
shape=shape,
|
||||||
|
strides=strides,
|
||||||
|
data=(dlt.data, True),
|
||||||
|
offset=dlt.byte_offset,
|
||||||
|
typestr=typestr, )
|
||||||
|
|
||||||
|
|
||||||
|
class _Holder:
|
||||||
|
"""A wrapper that combines a pycapsule and array_interface for consumption by numpy.
|
||||||
|
|
||||||
|
Parameters
|
||||||
|
----------
|
||||||
|
array_interface : dict
|
||||||
|
A description of the underlying memory.
|
||||||
|
|
||||||
|
pycapsule : PyCapsule
|
||||||
|
A wrapper around the dlpack tensor that will be converted to numpy.
|
||||||
|
"""
|
||||||
|
|
||||||
|
def __init__(self, array_interface, pycapsule) -> None:
|
||||||
|
self.__array_interface__ = array_interface
|
||||||
|
self._pycapsule = pycapsule
|
||||||
|
|
||||||
|
|
||||||
|
def to_numpy(pycapsule) -> np.ndarray:
|
||||||
|
"""Convert a dlpack tensor into a numpy array without copying.
|
||||||
|
|
||||||
|
Parameters
|
||||||
|
----------
|
||||||
|
pycapsule : PyCapsule
|
||||||
|
A pycapsule wrapping a dlpack tensor that will be converted.
|
||||||
|
|
||||||
|
Returns
|
||||||
|
-------
|
||||||
|
np_array : np.ndarray
|
||||||
|
A new numpy array that uses the same underlying memory as the input
|
||||||
|
pycapsule.
|
||||||
|
"""
|
||||||
|
assert ctypes.pythonapi.PyCapsule_IsValid(pycapsule, _c_str_dltensor)
|
||||||
|
dl_managed_tensor = ctypes.pythonapi.PyCapsule_GetPointer(pycapsule,
|
||||||
|
_c_str_dltensor)
|
||||||
|
dl_managed_tensor_ptr = ctypes.cast(dl_managed_tensor,
|
||||||
|
ctypes.POINTER(DLManagedTensor))
|
||||||
|
dl_managed_tensor = dl_managed_tensor_ptr.contents
|
||||||
|
holder = _Holder(
|
||||||
|
_array_interface_from_dl_tensor(dl_managed_tensor.dl_tensor),
|
||||||
|
pycapsule)
|
||||||
|
return np.ctypeslib.as_array(holder)
|
36
third_party/dlpack/apps/numpy_dlpack/test.py
vendored
Normal file
36
third_party/dlpack/apps/numpy_dlpack/test.py
vendored
Normal file
@@ -0,0 +1,36 @@
|
|||||||
|
import tvm
|
||||||
|
import numpy as np
|
||||||
|
from dlpack import from_numpy, to_numpy
|
||||||
|
|
||||||
|
|
||||||
|
def test_from_numpy():
|
||||||
|
"""Test the copy free conversion of numpy to a tvm ndarray."""
|
||||||
|
np_array = np.random.normal(size=[10, 10])
|
||||||
|
np_array_ref = np_array.copy()
|
||||||
|
tvm_array = tvm.nd.from_dlpack(from_numpy(np_array))
|
||||||
|
del np_array
|
||||||
|
np.testing.assert_equal(actual=tvm_array.numpy(), desired=np_array_ref)
|
||||||
|
del tvm_array
|
||||||
|
|
||||||
|
|
||||||
|
def test_to_numpy():
|
||||||
|
"""Test the copy free conversion of a tvm ndarray to a numpy array"""
|
||||||
|
tvm_array = tvm.nd.array(np.random.normal(size=[10, 10]))
|
||||||
|
np_array_ref = tvm_array.numpy()
|
||||||
|
np_array = to_numpy(tvm_array.__dlpack__())
|
||||||
|
del tvm_array
|
||||||
|
np.testing.assert_equal(actual=np_array, desired=np_array_ref)
|
||||||
|
del np_array
|
||||||
|
|
||||||
|
|
||||||
|
if __name__ == "__main__":
|
||||||
|
"""
|
||||||
|
Run both tests a bunch of times to make
|
||||||
|
sure the conversions and memory management are stable.
|
||||||
|
"""
|
||||||
|
print("### Testing from_numpy")
|
||||||
|
for i in range(10000):
|
||||||
|
test_from_numpy()
|
||||||
|
print("### Testing to_numpy")
|
||||||
|
for i in range(10000):
|
||||||
|
test_to_numpy()
|
38
third_party/dlpack/apps/numpy_dlpack/test_pure_numpy.py
vendored
Normal file
38
third_party/dlpack/apps/numpy_dlpack/test_pure_numpy.py
vendored
Normal file
@@ -0,0 +1,38 @@
|
|||||||
|
import numpy as np
|
||||||
|
from dlpack import from_numpy, to_numpy
|
||||||
|
|
||||||
|
|
||||||
|
def test_to_from_numpy_zero_copy():
|
||||||
|
"""Test the copy free conversion of numpy array via DLPack."""
|
||||||
|
np_ary = np.random.normal(size=[10, 10])
|
||||||
|
np_ary_big = np.random.normal(size=[12, 10])
|
||||||
|
dlpack_capsule = from_numpy(np_ary_big)
|
||||||
|
reconstructed_ary = to_numpy(dlpack_capsule)
|
||||||
|
del dlpack_capsule
|
||||||
|
np_ary_big[1:11] = np_ary
|
||||||
|
del np_ary_big
|
||||||
|
np.testing.assert_equal(actual=reconstructed_ary[1:11], desired=np_ary)
|
||||||
|
|
||||||
|
|
||||||
|
def test_to_from_numpy_memory():
|
||||||
|
"""Test that DLPack capsule keeps the source array alive"""
|
||||||
|
source_array = np.random.normal(size=[10, 10])
|
||||||
|
np_array_ref = source_array.copy()
|
||||||
|
dlpack_capsule = from_numpy(source_array)
|
||||||
|
del source_array
|
||||||
|
reconstructed_array = to_numpy(dlpack_capsule)
|
||||||
|
del dlpack_capsule
|
||||||
|
np.testing.assert_equal(actual=reconstructed_array, desired=np_array_ref)
|
||||||
|
|
||||||
|
|
||||||
|
if __name__ == "__main__":
|
||||||
|
"""
|
||||||
|
Run both tests a bunch of times to make
|
||||||
|
sure the conversions and memory management are stable.
|
||||||
|
"""
|
||||||
|
print("### Running `test_to_from_numpy_zero_copy`")
|
||||||
|
for i in range(10000):
|
||||||
|
test_to_from_numpy_zero_copy()
|
||||||
|
print("### Running `test_to_from_numpy_memory`")
|
||||||
|
for i in range(10000):
|
||||||
|
test_to_from_numpy_memory()
|
4
third_party/dlpack/cmake/template/Config.cmake.in
vendored
Normal file
4
third_party/dlpack/cmake/template/Config.cmake.in
vendored
Normal file
@@ -0,0 +1,4 @@
|
|||||||
|
@PACKAGE_INIT@
|
||||||
|
|
||||||
|
include("${CMAKE_CURRENT_LIST_DIR}/@TARGETS_EXPORT_NAME@.cmake")
|
||||||
|
check_required_components("@PROJECT_NAME@")
|
65
third_party/dlpack/contrib/dlpack/dlpackcpp.h
vendored
Normal file
65
third_party/dlpack/contrib/dlpack/dlpackcpp.h
vendored
Normal file
@@ -0,0 +1,65 @@
|
|||||||
|
/*!
|
||||||
|
* Copyright (c) 2017 by Contributors
|
||||||
|
* \file dlpackcpp.h
|
||||||
|
* \brief Example C++ wrapper of DLPack
|
||||||
|
*/
|
||||||
|
#ifndef DLPACK_DLPACKCPP_H_
|
||||||
|
#define DLPACK_DLPACKCPP_H_
|
||||||
|
|
||||||
|
#include <dlpack/dlpack.h>
|
||||||
|
|
||||||
|
#include <cstdint> // for int64_t etc
|
||||||
|
#include <cstdlib> // for free()
|
||||||
|
#include <functional> // for std::multiplies
|
||||||
|
#include <memory>
|
||||||
|
#include <numeric>
|
||||||
|
#include <vector>
|
||||||
|
|
||||||
|
namespace dlpack {
|
||||||
|
|
||||||
|
// Example container wrapping of DLTensor.
|
||||||
|
class DLTContainer {
|
||||||
|
public:
|
||||||
|
DLTContainer() {
|
||||||
|
// default to float32
|
||||||
|
handle_.data = nullptr;
|
||||||
|
handle_.dtype.code = kDLFloat;
|
||||||
|
handle_.dtype.bits = 32U;
|
||||||
|
handle_.dtype.lanes = 1U;
|
||||||
|
handle_.device.device_type = kDLCPU;
|
||||||
|
handle_.device.device_id = 0;
|
||||||
|
handle_.shape = nullptr;
|
||||||
|
handle_.strides = nullptr;
|
||||||
|
handle_.byte_offset = 0;
|
||||||
|
}
|
||||||
|
~DLTContainer() {
|
||||||
|
if (origin_ == nullptr) {
|
||||||
|
free(handle_.data);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
operator DLTensor() {
|
||||||
|
return handle_;
|
||||||
|
}
|
||||||
|
operator DLTensor*() {
|
||||||
|
return &(handle_);
|
||||||
|
}
|
||||||
|
void Reshape(const std::vector<int64_t>& shape) {
|
||||||
|
shape_ = shape;
|
||||||
|
int64_t sz = std::accumulate(std::begin(shape), std::end(shape),
|
||||||
|
int64_t(1), std::multiplies<int64_t>());
|
||||||
|
int ret = posix_memalign(&handle_.data, 256, sz);
|
||||||
|
if (ret != 0) throw std::bad_alloc();
|
||||||
|
handle_.shape = &shape_[0];
|
||||||
|
handle_.ndim = static_cast<uint32_t>(shape.size());
|
||||||
|
}
|
||||||
|
|
||||||
|
private:
|
||||||
|
DLTensor handle_;
|
||||||
|
std::vector<int64_t> shape_;
|
||||||
|
std::vector<int64_t> strides_;
|
||||||
|
// original space container, if
|
||||||
|
std::shared_ptr<DLTContainer> origin_;
|
||||||
|
};
|
||||||
|
|
||||||
|
} // namespace dlpack
|
||||||
|
#endif // DLPACK_DLPACKCPP_H_
|
7
third_party/dlpack/contrib/mock_c.c
vendored
Normal file
7
third_party/dlpack/contrib/mock_c.c
vendored
Normal file
@@ -0,0 +1,7 @@
|
|||||||
|
// Copyright by contributors
|
||||||
|
// This file is used to make sure the package is C compatible
|
||||||
|
#include <dlpack/dlpack.h>
|
||||||
|
|
||||||
|
int GetNDim(DLTensor *t) {
|
||||||
|
return t->ndim;
|
||||||
|
}
|
8
third_party/dlpack/contrib/mock_main.cc
vendored
Normal file
8
third_party/dlpack/contrib/mock_main.cc
vendored
Normal file
@@ -0,0 +1,8 @@
|
|||||||
|
// Copyright by contributors
|
||||||
|
#include <dlpack/dlpack.h>
|
||||||
|
#include <dlpack/dlpackcpp.h>
|
||||||
|
|
||||||
|
int main() {
|
||||||
|
dlpack::DLTContainer c;
|
||||||
|
return 0;
|
||||||
|
}
|
3
third_party/dlpack/doc_requirements.txt
vendored
Normal file
3
third_party/dlpack/doc_requirements.txt
vendored
Normal file
@@ -0,0 +1,3 @@
|
|||||||
|
Sphinx==4.4.0
|
||||||
|
pydata-sphinx-theme==0.7.1
|
||||||
|
breathe==4.31.0
|
1
third_party/dlpack/docs/.gitignore
vendored
Normal file
1
third_party/dlpack/docs/.gitignore
vendored
Normal file
@@ -0,0 +1 @@
|
|||||||
|
doxygen
|
37
third_party/dlpack/docs/CMakeLists.txt
vendored
Normal file
37
third_party/dlpack/docs/CMakeLists.txt
vendored
Normal file
@@ -0,0 +1,37 @@
|
|||||||
|
find_package(Doxygen QUIET)
|
||||||
|
if(NOT DOXYGEN_FOUND)
|
||||||
|
message(FATAL_ERROR "Doxygen is needed to build the documentation.")
|
||||||
|
endif()
|
||||||
|
|
||||||
|
# TODO: add config file
|
||||||
|
set(doxyfile_in ${CMAKE_CURRENT_SOURCE_DIR}/Doxyfile.in)
|
||||||
|
set(doxyfile ${CMAKE_CURRENT_BINARY_DIR}/Doxyfile)
|
||||||
|
set(doxygen_output_dir ${CMAKE_CURRENT_BINARY_DIR}/doxygen)
|
||||||
|
|
||||||
|
configure_file(${doxyfile_in} ${doxyfile} @ONLY)
|
||||||
|
|
||||||
|
file(MAKE_DIRECTORY ${doxygen_output_dir})
|
||||||
|
|
||||||
|
add_custom_target(Doxygen ALL
|
||||||
|
COMMAND ${DOXYGEN_EXECUTABLE} ${doxyfile}
|
||||||
|
WORKING_DIRECTORY ${CMAKE_CURRENT_BINARY_DIR}
|
||||||
|
COMMENT "Generating API documentation with Doxygen"
|
||||||
|
VERBATIM)
|
||||||
|
|
||||||
|
find_program(SPHINX_EXECUTABLE
|
||||||
|
NAMES sphinx-build
|
||||||
|
DOC "Path to sphinx-build executable")
|
||||||
|
|
||||||
|
set(sphinx_source ${CMAKE_CURRENT_SOURCE_DIR}/source)
|
||||||
|
set(sphinx_build ${CMAKE_CURRENT_BINARY_DIR}/build/latest)
|
||||||
|
set(doxygen_xml_builddir ${doxygen_output_dir}/xml)
|
||||||
|
|
||||||
|
add_custom_target(Sphinx ALL
|
||||||
|
COMMAND ${SPHINX_EXECUTABLE} -b html
|
||||||
|
-Dbreathe_projects.dlpack=${doxygen_xml_builddir}
|
||||||
|
${sphinx_source} ${sphinx_build} -WT --keep-going
|
||||||
|
WORKING_DIRECTORY ${CMAKE_CURRENT_BINARY_DIR}
|
||||||
|
COMMENT "Generating documentation with Sphinx"
|
||||||
|
VERBATIM)
|
||||||
|
|
||||||
|
install(DIRECTORY ${sphinx_build} DESTINATION share/${PROJECT_NAME}/docs)
|
2339
third_party/dlpack/docs/Doxyfile
vendored
Normal file
2339
third_party/dlpack/docs/Doxyfile
vendored
Normal file
File diff suppressed because it is too large
Load Diff
18
third_party/dlpack/docs/Doxyfile.in
vendored
Normal file
18
third_party/dlpack/docs/Doxyfile.in
vendored
Normal file
@@ -0,0 +1,18 @@
|
|||||||
|
PROJECT_NAME = @PROJECT_NAME@
|
||||||
|
PROJECT_NUMBER = @PROJECT_VERSION@
|
||||||
|
PROJECT_BRIEF = "Common in-memory tensor structure and operator interface for deep learning and other systems"
|
||||||
|
STRIP_FROM_PATH = @PROJECT_SOURCE_DIR@ \
|
||||||
|
@PROJECT_BINARY_DIR@
|
||||||
|
OUTPUT_LANGUAGE = English
|
||||||
|
FILE_PATTERNS = *.h *.md
|
||||||
|
RECURSIVE = YES
|
||||||
|
INPUT = @CMAKE_SOURCE_DIR@/include/dlpack
|
||||||
|
IMAGE_PATH = @CMAKE_SOURCE_DIR@/docs
|
||||||
|
USE_MDFILE_AS_MAINPAGE = @CMAKE_SOURCE_DIR@/docs/readme.md
|
||||||
|
JAVADOC_AUTOBRIEF = YES
|
||||||
|
GENERATE_HTML = NO
|
||||||
|
GENERATE_LATEX = NO
|
||||||
|
GENERATE_XML = YES
|
||||||
|
XML_OUTPUT = xml
|
||||||
|
XML_PROGRAMLISTING = YES
|
||||||
|
OUTPUT_DIRECTORY = @CMAKE_BINARY_DIR@/docs/doxygen
|
20
third_party/dlpack/docs/Makefile
vendored
Normal file
20
third_party/dlpack/docs/Makefile
vendored
Normal file
@@ -0,0 +1,20 @@
|
|||||||
|
# Minimal makefile for Sphinx documentation
|
||||||
|
#
|
||||||
|
|
||||||
|
# You can set these variables from the command line, and also
|
||||||
|
# from the environment for the first two.
|
||||||
|
SPHINXOPTS ?= -Dbreathe_projects.dlpack=../doxygen/xml -WT --keep-going
|
||||||
|
SPHINXBUILD ?= sphinx-build
|
||||||
|
SOURCEDIR = source
|
||||||
|
BUILDDIR = build/latest
|
||||||
|
|
||||||
|
# Put it first so that "make" without argument is like "make help".
|
||||||
|
help:
|
||||||
|
@$(SPHINXBUILD) -M help "$(SOURCEDIR)" "$(BUILDDIR)" $(SPHINXOPTS) $(O)
|
||||||
|
|
||||||
|
.PHONY: help Makefile
|
||||||
|
|
||||||
|
# Catch-all target: route all unknown targets to Sphinx using the new
|
||||||
|
# "make mode" option. $(O) is meant as a shortcut for $(SPHINXOPTS).
|
||||||
|
%: Makefile
|
||||||
|
@$(SPHINXBUILD) -b $@ "$(SOURCEDIR)" "$(BUILDDIR)" $(SPHINXOPTS) $(O)
|
13
third_party/dlpack/docs/README.md
vendored
Normal file
13
third_party/dlpack/docs/README.md
vendored
Normal file
@@ -0,0 +1,13 @@
|
|||||||
|
# Documentation for DLPack
|
||||||
|
|
||||||
|
## Building Locally
|
||||||
|
|
||||||
|
The following dependencies must be downloaded in order to build docs:
|
||||||
|
|
||||||
|
- Doxygen (On debian distros, simply run `sudo apt -y install doxygen`)
|
||||||
|
- Python dependencies present in the `doc_requirements.txt` file in the root directory of the project. Run `python3 -m pip install -r doc_requirements.txt` to install them.
|
||||||
|
|
||||||
|
Once the dependencies are installed, docs can be built using either CMake or the Makefile from the root directory of the project.
|
||||||
|
|
||||||
|
- Using Makefile: Run `make doc` to build the HTML pages. Run `make show_docs` to serve the website locally.
|
||||||
|
- Using CMake: Build with `BUILD_DOCS` option `ON`: `mkdir -p build && cd build && cmake .. -DBUILD_DOCS=ON && make`
|
36
third_party/dlpack/docs/make.bat
vendored
Normal file
36
third_party/dlpack/docs/make.bat
vendored
Normal file
@@ -0,0 +1,36 @@
|
|||||||
|
@ECHO OFF
|
||||||
|
|
||||||
|
pushd %~dp0
|
||||||
|
|
||||||
|
REM Command file for Sphinx documentation
|
||||||
|
|
||||||
|
if "%SPHINXBUILD%" == "" (
|
||||||
|
set SPHINXBUILD=sphinx-build
|
||||||
|
)
|
||||||
|
set SOURCEDIR=source
|
||||||
|
set SPHINXOPTS=-Dbreathe_projects.dlpack=../doxygen/xml -WT --keep-going
|
||||||
|
set BUILDDIR=build/latest
|
||||||
|
|
||||||
|
if "%1" == "" goto help
|
||||||
|
|
||||||
|
%SPHINXBUILD% >NUL 2>NUL
|
||||||
|
if errorlevel 9009 (
|
||||||
|
echo.
|
||||||
|
echo.The 'sphinx-build' command was not found. Make sure you have Sphinx
|
||||||
|
echo.installed, then set the SPHINXBUILD environment variable to point
|
||||||
|
echo.to the full path of the 'sphinx-build' executable. Alternatively you
|
||||||
|
echo.may add the Sphinx directory to PATH.
|
||||||
|
echo.
|
||||||
|
echo.If you don't have Sphinx installed, grab it from
|
||||||
|
echo.https://www.sphinx-doc.org/
|
||||||
|
exit /b 1
|
||||||
|
)
|
||||||
|
|
||||||
|
%SPHINXBUILD% -b %1 %SOURCEDIR% %BUILDDIR% %SPHINXOPTS% %O%
|
||||||
|
goto end
|
||||||
|
|
||||||
|
:help
|
||||||
|
%SPHINXBUILD% -M help %SOURCEDIR% %BUILDDIR% %SPHINXOPTS% %O%
|
||||||
|
|
||||||
|
:end
|
||||||
|
popd
|
BIN
third_party/dlpack/docs/source/_static/images/DLPack_diagram.png
vendored
Normal file
BIN
third_party/dlpack/docs/source/_static/images/DLPack_diagram.png
vendored
Normal file
Binary file not shown.
After Width: | Height: | Size: 23 KiB |
35
third_party/dlpack/docs/source/c_api.rst
vendored
Normal file
35
third_party/dlpack/docs/source/c_api.rst
vendored
Normal file
@@ -0,0 +1,35 @@
|
|||||||
|
.. _c_api:
|
||||||
|
|
||||||
|
C API (``dlpack.h``)
|
||||||
|
====================
|
||||||
|
|
||||||
|
Macros
|
||||||
|
~~~~~~
|
||||||
|
|
||||||
|
.. doxygendefine:: DLPACK_EXTERN_C
|
||||||
|
|
||||||
|
.. doxygendefine:: DLPACK_VERSION
|
||||||
|
|
||||||
|
.. doxygendefine:: DLPACK_DLL
|
||||||
|
|
||||||
|
Enumerations
|
||||||
|
~~~~~~~~~~~~
|
||||||
|
|
||||||
|
.. doxygenenum:: DLDeviceType
|
||||||
|
|
||||||
|
.. doxygenenum:: DLDataTypeCode
|
||||||
|
|
||||||
|
Structs
|
||||||
|
~~~~~~~
|
||||||
|
|
||||||
|
.. doxygenstruct:: DLDevice
|
||||||
|
:members:
|
||||||
|
|
||||||
|
.. doxygenstruct:: DLDataType
|
||||||
|
:members:
|
||||||
|
|
||||||
|
.. doxygenstruct:: DLTensor
|
||||||
|
:members:
|
||||||
|
|
||||||
|
.. doxygenstruct:: DLManagedTensor
|
||||||
|
:members:
|
78
third_party/dlpack/docs/source/conf.py
vendored
Normal file
78
third_party/dlpack/docs/source/conf.py
vendored
Normal file
@@ -0,0 +1,78 @@
|
|||||||
|
# Configuration file for the Sphinx documentation builder.
|
||||||
|
#
|
||||||
|
# This file only contains a selection of the most common options. For a full
|
||||||
|
# list see the documentation:
|
||||||
|
# https://www.sphinx-doc.org/en/master/usage/configuration.html
|
||||||
|
|
||||||
|
# -- Path setup --------------------------------------------------------------
|
||||||
|
|
||||||
|
# If extensions (or modules to document with autodoc) are in another directory,
|
||||||
|
# add these directories to sys.path here. If the directory is relative to the
|
||||||
|
# documentation root, use os.path.abspath to make it absolute, like shown here.
|
||||||
|
#
|
||||||
|
# import os
|
||||||
|
# import sys
|
||||||
|
# sys.path.insert(0, os.path.abspath('.'))
|
||||||
|
|
||||||
|
# -- Project information -----------------------------------------------------
|
||||||
|
|
||||||
|
project = 'DLPack'
|
||||||
|
copyright = '2022, DLPack contributors'
|
||||||
|
author = 'DLPack contributors'
|
||||||
|
|
||||||
|
# The full version, including alpha/beta/rc tags
|
||||||
|
release = '0.6.0'
|
||||||
|
|
||||||
|
# -- General configuration ---------------------------------------------------
|
||||||
|
|
||||||
|
# Add any Sphinx extension module names here, as strings. They can be
|
||||||
|
# extensions coming with Sphinx (named 'sphinx.ext.*') or your custom
|
||||||
|
# ones.
|
||||||
|
extensions = ['sphinx.ext.intersphinx', 'breathe']
|
||||||
|
|
||||||
|
# Add any paths that contain templates here, relative to this directory.
|
||||||
|
# templates_path = ['_templates']
|
||||||
|
|
||||||
|
# List of patterns, relative to source directory, that match files and
|
||||||
|
# directories to ignore when looking for source files.
|
||||||
|
# This pattern also affects html_static_path and html_extra_path.
|
||||||
|
exclude_patterns = []
|
||||||
|
|
||||||
|
intersphinx_mapping = {
|
||||||
|
"array_api": ("https://data-apis.org/array-api/latest", None),
|
||||||
|
}
|
||||||
|
|
||||||
|
html_use_index = True
|
||||||
|
html_domain_indices = True
|
||||||
|
|
||||||
|
html_theme_options = {
|
||||||
|
"github_url": "https://github.com/dmlc/dlpack",
|
||||||
|
"use_edit_page_button": True,
|
||||||
|
"show_toc_level": 1,
|
||||||
|
}
|
||||||
|
|
||||||
|
html_context = {
|
||||||
|
"github_user": "dmlc",
|
||||||
|
"github_repo": "dlpack",
|
||||||
|
"github_version": "main",
|
||||||
|
"doc_path": "docs/source",
|
||||||
|
}
|
||||||
|
|
||||||
|
# -- Breathe configuration ---------------------------------------------------
|
||||||
|
|
||||||
|
# Tell breathe what to call the project. This is used to link XML files
|
||||||
|
# generated by Doxygen by passing a command line option:
|
||||||
|
# -Dbreathe_projects.<breathe_default_project>=</path/to/xml/index/directory>
|
||||||
|
breathe_default_project = "dlpack"
|
||||||
|
|
||||||
|
# -- Options for HTML output -------------------------------------------------
|
||||||
|
|
||||||
|
# The theme to use for HTML and HTML Help pages. See the documentation for
|
||||||
|
# a list of builtin themes.
|
||||||
|
#
|
||||||
|
html_theme = 'pydata_sphinx_theme' # 'alabaster'
|
||||||
|
|
||||||
|
# Add any paths that contain custom static files (such as style sheets) here,
|
||||||
|
# relative to this directory. They are copied after the builtin static files,
|
||||||
|
# so a file named "default.css" will overwrite the builtin "default.css".
|
||||||
|
html_static_path = ['_static']
|
82
third_party/dlpack/docs/source/index.rst
vendored
Normal file
82
third_party/dlpack/docs/source/index.rst
vendored
Normal file
@@ -0,0 +1,82 @@
|
|||||||
|
Welcome to DLPack's documentation!
|
||||||
|
==================================
|
||||||
|
|
||||||
|
|
||||||
|
Purpose
|
||||||
|
~~~~~~~
|
||||||
|
|
||||||
|
In order for an ndarray system to interact with a variety of frameworks, a
|
||||||
|
stable in-memory data structure is needed.
|
||||||
|
|
||||||
|
DLPack is one such data structure that allows exchange between major
|
||||||
|
frameworks. It is developed with inputs from many deep learning system core
|
||||||
|
developers. Highlights include:
|
||||||
|
|
||||||
|
* Minimum and stable: :ref:`simple header <c_api>`
|
||||||
|
* Designed for cross hardware: CPU, CUDA, OpenCL, Vulkan, Metal, VPI, ROCm,
|
||||||
|
WebGPU, Hexagon
|
||||||
|
* Already a standard with wide community adoption and support:
|
||||||
|
|
||||||
|
* `NumPy <https://numpy.org/doc/stable/release/1.22.0-notes.html#add-nep-47-compatible-dlpack-support>`_
|
||||||
|
* `CuPy <https://docs.cupy.dev/en/stable/reference/generated/cupy.fromDlpack.html>`_
|
||||||
|
* `PyTorch <https://pytorch.org/docs/stable/dlpack.html>`_
|
||||||
|
* `Tensorflow <https://www.tensorflow.org/api_docs/python/tf/experimental/dlpack/from_dlpack>`_
|
||||||
|
* `MXNet <https://mxnet.apache.org/versions/master/api/python/docs/_modules/mxnet/dlpack.html>`_
|
||||||
|
* `TVM <https://tvm.apache.org/docs/reference/api/python/contrib.html#module-tvm.contrib.dlpack>`_
|
||||||
|
* `mpi4py <https://mpi4py.readthedocs.io/en/stable/overview.html#support-for-gpu-aware-mpi>`_
|
||||||
|
|
||||||
|
* Clean C ABI compatible.
|
||||||
|
|
||||||
|
* Means you can create and access it from any language.
|
||||||
|
* It is also essential for building JIT and AOT compilers to support these
|
||||||
|
data types.
|
||||||
|
|
||||||
|
|
||||||
|
Scope
|
||||||
|
~~~~~
|
||||||
|
|
||||||
|
The main design rationale of DLPack is the minimalism. DLPack drops the
|
||||||
|
consideration of allocator, device API and focus on the minimum data
|
||||||
|
structure. While still considering the need for cross hardware support
|
||||||
|
(e.g. the data field is opaque for platforms that does not support normal
|
||||||
|
addressing).
|
||||||
|
|
||||||
|
It also simplifies some of the design to remove legacy issues (e.g. everything
|
||||||
|
assumes to be row major, strides can be used to support other case, and avoid
|
||||||
|
the complexity to consider more layouts).
|
||||||
|
|
||||||
|
|
||||||
|
Roadmap
|
||||||
|
~~~~~~~
|
||||||
|
|
||||||
|
* C API that could be exposed as a new Python attribute ``__dlpack_info__``
|
||||||
|
for returning API and ABI versions. (see `#34 <https://github.com/dmlc/dlpack/issues/34>`_,
|
||||||
|
`#72 <https://github.com/dmlc/dlpack/pull/72>`_)
|
||||||
|
* Clarify alignment requirements. (see
|
||||||
|
`data-apis/array-api#293 <https://github.com/data-apis/array-api/issues/293>`_,
|
||||||
|
`numpy/numpy#20338 <https://github.com/numpy/numpy/issues/20338>`_,
|
||||||
|
`data-apis/array-api#293 (comment) <https://github.com/data-apis/array-api/issues/293#issuecomment-964434449>`_)
|
||||||
|
* Adding support for boolean data type (see `#75 <https://github.com/dmlc/dlpack/issues/75>`_)
|
||||||
|
* Adding a read-only flag (ABI break) or making it a hard requirement in the spec that
|
||||||
|
imported arrays should be treated as read-only. (see
|
||||||
|
`data-apis/consortium-feedback#1 (comment) <https://github.com/data-apis/consortium-feedback/issues/1#issuecomment-675857753>`_,
|
||||||
|
`data-apis/array-api#191 <https://github.com/data-apis/array-api/issues/191>`_)
|
||||||
|
* Standardize C interface for stream exchange. (see `#74 <https://github.com/dmlc/dlpack/issues/74>`_,
|
||||||
|
`#65 <https://github.com/dmlc/dlpack/issues/65>`_)
|
||||||
|
|
||||||
|
|
||||||
|
DLPack Documentation
|
||||||
|
~~~~~~~~~~~~~~~~~~~~
|
||||||
|
|
||||||
|
.. toctree::
|
||||||
|
:maxdepth: 2
|
||||||
|
|
||||||
|
c_api
|
||||||
|
python_spec
|
||||||
|
|
||||||
|
|
||||||
|
Indices and tables
|
||||||
|
==================
|
||||||
|
|
||||||
|
* :ref:`genindex`
|
||||||
|
* :ref:`search`
|
156
third_party/dlpack/docs/source/python_spec.rst
vendored
Normal file
156
third_party/dlpack/docs/source/python_spec.rst
vendored
Normal file
@@ -0,0 +1,156 @@
|
|||||||
|
.. _python-spec:
|
||||||
|
|
||||||
|
Python Specification for DLPack
|
||||||
|
===============================
|
||||||
|
|
||||||
|
The Python specification for DLPack is a part of the
|
||||||
|
`Python array API standard <https://data-apis.org/array-api/latest/index.html>`_.
|
||||||
|
More details about the spec can be found under the :ref:`data-interchange` page.
|
||||||
|
|
||||||
|
|
||||||
|
Syntax for data interchange with DLPack
|
||||||
|
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
|
||||||
|
|
||||||
|
The array API will offer the following syntax for data interchange:
|
||||||
|
|
||||||
|
1. A ``from_dlpack(x)`` function, which accepts (array) objects with a
|
||||||
|
``__dlpack__`` method and uses that method to construct a new array
|
||||||
|
containing the data from ``x``.
|
||||||
|
2. ``__dlpack__(self, stream=None)`` and ``__dlpack_device__`` methods on the
|
||||||
|
array object, which will be called from within ``from_dlpack``, to query
|
||||||
|
what device the array is on (may be needed to pass in the correct
|
||||||
|
stream, e.g. in the case of multiple GPUs) and to access the data.
|
||||||
|
|
||||||
|
|
||||||
|
Semantics
|
||||||
|
~~~~~~~~~
|
||||||
|
|
||||||
|
DLPack describes the memory layout of strided, n-dimensional arrays.
|
||||||
|
When a user calls ``y = from_dlpack(x)``, the library implementing ``x`` (the
|
||||||
|
"producer") will provide access to the data from ``x`` to the library
|
||||||
|
containing ``from_dlpack`` (the "consumer"). If possible, this must be
|
||||||
|
zero-copy (i.e. ``y`` will be a *view* on ``x``). If not possible, that library
|
||||||
|
may make a copy of the data. In both cases:
|
||||||
|
|
||||||
|
- The producer keeps owning the memory
|
||||||
|
- ``y`` may or may not be a view, therefore the user must keep the recommendation to
|
||||||
|
avoid mutating ``y`` in mind - see :ref:`copyview-mutability`.
|
||||||
|
- Both ``x`` and ``y`` may continue to be used just like arrays created in other ways.
|
||||||
|
|
||||||
|
If an array that is accessed via the interchange protocol lives on a
|
||||||
|
device that the requesting library does not support, it is recommended to
|
||||||
|
raise a ``TypeError``.
|
||||||
|
|
||||||
|
Stream handling through the ``stream`` keyword applies to CUDA and ROCm (perhaps
|
||||||
|
to other devices that have a stream concept as well, however those haven't been
|
||||||
|
considered in detail). The consumer must pass the stream it will use to the
|
||||||
|
producer; the producer must synchronize or wait on the stream when necessary.
|
||||||
|
In the common case of the default stream being used, synchronization will be
|
||||||
|
unnecessary so asynchronous execution is enabled.
|
||||||
|
|
||||||
|
|
||||||
|
Implementation
|
||||||
|
~~~~~~~~~~~~~~
|
||||||
|
|
||||||
|
*Note that while this API standard largely tries to avoid discussing
|
||||||
|
implementation details, some discussion and requirements are needed
|
||||||
|
here because data interchange requires coordination between
|
||||||
|
implementers on, e.g., memory management.*
|
||||||
|
|
||||||
|
.. image:: /_static/images/DLPack_diagram.png
|
||||||
|
:alt: Diagram of DLPack structs
|
||||||
|
|
||||||
|
*DLPack diagram. Dark blue are the structs it defines, light blue
|
||||||
|
struct members, gray text enum values of supported devices and data
|
||||||
|
types.*
|
||||||
|
|
||||||
|
The ``__dlpack__`` method will produce a ``PyCapsule`` containing a
|
||||||
|
``DLManagedTensor``, which will be consumed immediately within
|
||||||
|
``from_dlpack`` - therefore it is consumed exactly once, and it will not be
|
||||||
|
visible to users of the Python API.
|
||||||
|
|
||||||
|
The producer must set the ``PyCapsule`` name to ``"dltensor"`` so that
|
||||||
|
it can be inspected by name, and set ``PyCapsule_Destructor`` that calls
|
||||||
|
the ``deleter`` of the ``DLManagedTensor`` when the ``"dltensor"``-named
|
||||||
|
capsule is no longer needed.
|
||||||
|
|
||||||
|
The consumer must transer ownership of the ``DLManangedTensor`` from the
|
||||||
|
capsule to its own object. It does so by renaming the capsule to
|
||||||
|
``"used_dltensor"`` to ensure that ``PyCapsule_Destructor`` will not get
|
||||||
|
called (ensured if ``PyCapsule_Destructor`` calls ``deleter`` only for
|
||||||
|
capsules whose name is ``"dltensor"``), but the ``deleter`` of the
|
||||||
|
``DLManagedTensor`` will be called by the destructor of the consumer
|
||||||
|
library object created to own the ``DLManagerTensor`` obtained from the
|
||||||
|
capsule. Below is an example of the capsule deleter written in the Python
|
||||||
|
C API which is called either when the refcount on the capsule named
|
||||||
|
``"dltensor"`` reaches zero or the consumer decides to deallocate its array:
|
||||||
|
|
||||||
|
.. code-block:: C
|
||||||
|
|
||||||
|
static void dlpack_capsule_deleter(PyObject *self){
|
||||||
|
if (PyCapsule_IsValid(self, "used_dltensor")) {
|
||||||
|
return; /* Do nothing if the capsule has been consumed. */
|
||||||
|
}
|
||||||
|
|
||||||
|
/* an exception may be in-flight, we must save it in case we create another one */
|
||||||
|
PyObject *type, *value, *traceback;
|
||||||
|
PyErr_Fetch(&type, &value, &traceback);
|
||||||
|
|
||||||
|
DLManagedTensor *managed = (DLManagedTensor *)PyCapsule_GetPointer(self, "dltensor");
|
||||||
|
if (managed == NULL) {
|
||||||
|
PyErr_WriteUnraisable(self);
|
||||||
|
goto done;
|
||||||
|
}
|
||||||
|
/* the spec says the deleter can be NULL if there is no way for the caller to provide a reasonable destructor. */
|
||||||
|
if (managed->deleter) {
|
||||||
|
managed->deleter(managed);
|
||||||
|
/* TODO: is the deleter allowed to set a python exception? */
|
||||||
|
assert(!PyErr_Occurred());
|
||||||
|
}
|
||||||
|
|
||||||
|
done:
|
||||||
|
PyErr_Restore(type, value, traceback);
|
||||||
|
}
|
||||||
|
|
||||||
|
Note: the capsule names ``"dltensor"`` and ``"used_dltensor"`` must be
|
||||||
|
statically allocated.
|
||||||
|
|
||||||
|
When the ``strides`` field in the ``DLTensor`` struct is ``NULL``, it indicates a
|
||||||
|
row-major compact array. If the array is of size zero, the data pointer in
|
||||||
|
``DLTensor`` should be set to either ``NULL`` or ``0``.
|
||||||
|
|
||||||
|
DLPack version used must be ``0.2 <= DLPACK_VERSION < 1.0``. For further
|
||||||
|
details on DLPack design and how to implement support for it,
|
||||||
|
refer to `github.com/dmlc/dlpack <https://github.com/dmlc/dlpack>`_.
|
||||||
|
|
||||||
|
.. warning::
|
||||||
|
DLPack contains a ``device_id``, which will be the device
|
||||||
|
ID (an integer, ``0, 1, ...``) which the producer library uses. In
|
||||||
|
practice this will likely be the same numbering as that of the
|
||||||
|
consumer, however that is not guaranteed. Depending on the hardware
|
||||||
|
type, it may be possible for the consumer library implementation to
|
||||||
|
look up the actual device from the pointer to the data - this is
|
||||||
|
possible for example for CUDA device pointers.
|
||||||
|
|
||||||
|
It is recommended that implementers of this array API consider and document
|
||||||
|
whether the ``.device`` attribute of the array returned from ``from_dlpack`` is
|
||||||
|
guaranteed to be in a certain order or not.
|
||||||
|
|
||||||
|
|
||||||
|
Reference Implementations
|
||||||
|
~~~~~~~~~~~~~~~~~~~~~~~~~
|
||||||
|
|
||||||
|
Several Python libraries have adopted this standard using Python C API, C++, Cython,
|
||||||
|
ctypes, cffi, etc:
|
||||||
|
|
||||||
|
* NumPy: `Python C API <https://github.com/numpy/numpy/blob/main/numpy/core/src/multiarray/dlpack.c>`__
|
||||||
|
* CuPy: `Cython <https://github.com/cupy/cupy/blob/master/cupy/_core/dlpack.pyx>`__
|
||||||
|
* Tensorflow: `C++ <https://github.com/tensorflow/tensorflow/blob/master/tensorflow/c/eager/dlpack.cc>`__,
|
||||||
|
`Python wrapper using Python C API <https://github.com/tensorflow/tensorflow/blob/a97b01a4ff009ed84a571c138837130a311e74a7/tensorflow/python/tfe_wrapper.cc#L1562>`__,
|
||||||
|
`XLA <https://github.com/tensorflow/tensorflow/blob/master/tensorflow/compiler/xla/python/dlpack.cc>`__
|
||||||
|
* PyTorch: `C++ <https://github.com/pytorch/pytorch/blob/master/aten/src/ATen/DLConvertor.cpp>`__,
|
||||||
|
`Python wrapper using Python C API <https://github.com/pytorch/pytorch/blob/c22b8a42e6038ed2f6a161114cf3d8faac3f6e9a/torch/csrc/Module.cpp#L355>`__
|
||||||
|
* MXNet: `ctypes <https://github.com/apache/incubator-mxnet/blob/master/python/mxnet/dlpack.py>`__
|
||||||
|
* TVM: `ctypes <https://github.com/apache/tvm/blob/main/python/tvm/_ffi/_ctypes/ndarray.py>`__,
|
||||||
|
`Cython <https://github.com/apache/tvm/blob/main/python/tvm/_ffi/_cython/ndarray.pxi>`__
|
||||||
|
* mpi4py: `Cython <https://github.com/mpi4py/mpi4py/blob/master/src/mpi4py/MPI/asdlpack.pxi>`_
|
229
third_party/dlpack/include/dlpack/dlpack.h
vendored
Normal file
229
third_party/dlpack/include/dlpack/dlpack.h
vendored
Normal file
@@ -0,0 +1,229 @@
|
|||||||
|
/*!
|
||||||
|
* Copyright (c) 2017 by Contributors
|
||||||
|
* \file dlpack.h
|
||||||
|
* \brief The common header of DLPack.
|
||||||
|
*/
|
||||||
|
#ifndef DLPACK_DLPACK_H_
|
||||||
|
#define DLPACK_DLPACK_H_
|
||||||
|
|
||||||
|
/**
|
||||||
|
* \brief Compatibility with C++
|
||||||
|
*/
|
||||||
|
#ifdef __cplusplus
|
||||||
|
#define DLPACK_EXTERN_C extern "C"
|
||||||
|
#else
|
||||||
|
#define DLPACK_EXTERN_C
|
||||||
|
#endif
|
||||||
|
|
||||||
|
/*! \brief The current version of dlpack */
|
||||||
|
#define DLPACK_VERSION 70
|
||||||
|
|
||||||
|
/*! \brief The current ABI version of dlpack */
|
||||||
|
#define DLPACK_ABI_VERSION 1
|
||||||
|
|
||||||
|
/*! \brief DLPACK_DLL prefix for windows */
|
||||||
|
#ifdef _WIN32
|
||||||
|
#ifdef DLPACK_EXPORTS
|
||||||
|
#define DLPACK_DLL __declspec(dllexport)
|
||||||
|
#else
|
||||||
|
#define DLPACK_DLL __declspec(dllimport)
|
||||||
|
#endif
|
||||||
|
#else
|
||||||
|
#define DLPACK_DLL
|
||||||
|
#endif
|
||||||
|
|
||||||
|
#include <stdint.h>
|
||||||
|
#include <stddef.h>
|
||||||
|
|
||||||
|
#ifdef __cplusplus
|
||||||
|
extern "C" {
|
||||||
|
#endif
|
||||||
|
/*!
|
||||||
|
* \brief The device type in DLDevice.
|
||||||
|
*/
|
||||||
|
#ifdef __cplusplus
|
||||||
|
typedef enum : int32_t {
|
||||||
|
#else
|
||||||
|
typedef enum {
|
||||||
|
#endif
|
||||||
|
/*! \brief CPU device */
|
||||||
|
kDLCPU = 1,
|
||||||
|
/*! \brief CUDA GPU device */
|
||||||
|
kDLCUDA = 2,
|
||||||
|
/*!
|
||||||
|
* \brief Pinned CUDA CPU memory by cudaMallocHost
|
||||||
|
*/
|
||||||
|
kDLCUDAHost = 3,
|
||||||
|
/*! \brief OpenCL devices. */
|
||||||
|
kDLOpenCL = 4,
|
||||||
|
/*! \brief Vulkan buffer for next generation graphics. */
|
||||||
|
kDLVulkan = 7,
|
||||||
|
/*! \brief Metal for Apple GPU. */
|
||||||
|
kDLMetal = 8,
|
||||||
|
/*! \brief Verilog simulator buffer */
|
||||||
|
kDLVPI = 9,
|
||||||
|
/*! \brief ROCm GPUs for AMD GPUs */
|
||||||
|
kDLROCM = 10,
|
||||||
|
/*!
|
||||||
|
* \brief Pinned ROCm CPU memory allocated by hipMallocHost
|
||||||
|
*/
|
||||||
|
kDLROCMHost = 11,
|
||||||
|
/*!
|
||||||
|
* \brief Reserved extension device type,
|
||||||
|
* used for quickly test extension device
|
||||||
|
* The semantics can differ depending on the implementation.
|
||||||
|
*/
|
||||||
|
kDLExtDev = 12,
|
||||||
|
/*!
|
||||||
|
* \brief CUDA managed/unified memory allocated by cudaMallocManaged
|
||||||
|
*/
|
||||||
|
kDLCUDAManaged = 13,
|
||||||
|
/*!
|
||||||
|
* \brief Unified shared memory allocated on a oneAPI non-partititioned
|
||||||
|
* device. Call to oneAPI runtime is required to determine the device
|
||||||
|
* type, the USM allocation type and the sycl context it is bound to.
|
||||||
|
*
|
||||||
|
*/
|
||||||
|
kDLOneAPI = 14,
|
||||||
|
/*! \brief GPU support for next generation WebGPU standard. */
|
||||||
|
kDLWebGPU = 15,
|
||||||
|
/*! \brief Qualcomm Hexagon DSP */
|
||||||
|
kDLHexagon = 16,
|
||||||
|
} DLDeviceType;
|
||||||
|
|
||||||
|
/*!
|
||||||
|
* \brief A Device for Tensor and operator.
|
||||||
|
*/
|
||||||
|
typedef struct {
|
||||||
|
/*! \brief The device type used in the device. */
|
||||||
|
DLDeviceType device_type;
|
||||||
|
/*!
|
||||||
|
* \brief The device index.
|
||||||
|
* For vanilla CPU memory, pinned memory, or managed memory, this is set to 0.
|
||||||
|
*/
|
||||||
|
int32_t device_id;
|
||||||
|
} DLDevice;
|
||||||
|
|
||||||
|
/*!
|
||||||
|
* \brief The type code options DLDataType.
|
||||||
|
*/
|
||||||
|
typedef enum {
|
||||||
|
/*! \brief signed integer */
|
||||||
|
kDLInt = 0U,
|
||||||
|
/*! \brief unsigned integer */
|
||||||
|
kDLUInt = 1U,
|
||||||
|
/*! \brief IEEE floating point */
|
||||||
|
kDLFloat = 2U,
|
||||||
|
/*!
|
||||||
|
* \brief Opaque handle type, reserved for testing purposes.
|
||||||
|
* Frameworks need to agree on the handle data type for the exchange to be well-defined.
|
||||||
|
*/
|
||||||
|
kDLOpaqueHandle = 3U,
|
||||||
|
/*! \brief bfloat16 */
|
||||||
|
kDLBfloat = 4U,
|
||||||
|
/*!
|
||||||
|
* \brief complex number
|
||||||
|
* (C/C++/Python layout: compact struct per complex number)
|
||||||
|
*/
|
||||||
|
kDLComplex = 5U,
|
||||||
|
} DLDataTypeCode;
|
||||||
|
|
||||||
|
/*!
|
||||||
|
* \brief The data type the tensor can hold. The data type is assumed to follow the
|
||||||
|
* native endian-ness. An explicit error message should be raised when attempting to
|
||||||
|
* export an array with non-native endianness
|
||||||
|
*
|
||||||
|
* Examples
|
||||||
|
* - float: type_code = 2, bits = 32, lanes=1
|
||||||
|
* - float4(vectorized 4 float): type_code = 2, bits = 32, lanes=4
|
||||||
|
* - int8: type_code = 0, bits = 8, lanes=1
|
||||||
|
* - std::complex<float>: type_code = 5, bits = 64, lanes = 1
|
||||||
|
*/
|
||||||
|
typedef struct {
|
||||||
|
/*!
|
||||||
|
* \brief Type code of base types.
|
||||||
|
* We keep it uint8_t instead of DLDataTypeCode for minimal memory
|
||||||
|
* footprint, but the value should be one of DLDataTypeCode enum values.
|
||||||
|
* */
|
||||||
|
uint8_t code;
|
||||||
|
/*!
|
||||||
|
* \brief Number of bits, common choices are 8, 16, 32.
|
||||||
|
*/
|
||||||
|
uint8_t bits;
|
||||||
|
/*! \brief Number of lanes in the type, used for vector types. */
|
||||||
|
uint16_t lanes;
|
||||||
|
} DLDataType;
|
||||||
|
|
||||||
|
/*!
|
||||||
|
* \brief Plain C Tensor object, does not manage memory.
|
||||||
|
*/
|
||||||
|
typedef struct {
|
||||||
|
/*!
|
||||||
|
* \brief The data pointer points to the allocated data. This will be CUDA
|
||||||
|
* device pointer or cl_mem handle in OpenCL. It may be opaque on some device
|
||||||
|
* types. This pointer is always aligned to 256 bytes as in CUDA. The
|
||||||
|
* `byte_offset` field should be used to point to the beginning of the data.
|
||||||
|
*
|
||||||
|
* Note that as of Nov 2021, multiply libraries (CuPy, PyTorch, TensorFlow,
|
||||||
|
* TVM, perhaps others) do not adhere to this 256 byte aligment requirement
|
||||||
|
* on CPU/CUDA/ROCm, and always use `byte_offset=0`. This must be fixed
|
||||||
|
* (after which this note will be updated); at the moment it is recommended
|
||||||
|
* to not rely on the data pointer being correctly aligned.
|
||||||
|
*
|
||||||
|
* For given DLTensor, the size of memory required to store the contents of
|
||||||
|
* data is calculated as follows:
|
||||||
|
*
|
||||||
|
* \code{.c}
|
||||||
|
* static inline size_t GetDataSize(const DLTensor* t) {
|
||||||
|
* size_t size = 1;
|
||||||
|
* for (tvm_index_t i = 0; i < t->ndim; ++i) {
|
||||||
|
* size *= t->shape[i];
|
||||||
|
* }
|
||||||
|
* size *= (t->dtype.bits * t->dtype.lanes + 7) / 8;
|
||||||
|
* return size;
|
||||||
|
* }
|
||||||
|
* \endcode
|
||||||
|
*/
|
||||||
|
void* data;
|
||||||
|
/*! \brief The device of the tensor */
|
||||||
|
DLDevice device;
|
||||||
|
/*! \brief Number of dimensions */
|
||||||
|
int32_t ndim;
|
||||||
|
/*! \brief The data type of the pointer*/
|
||||||
|
DLDataType dtype;
|
||||||
|
/*! \brief The shape of the tensor */
|
||||||
|
int64_t* shape;
|
||||||
|
/*!
|
||||||
|
* \brief strides of the tensor (in number of elements, not bytes)
|
||||||
|
* can be NULL, indicating tensor is compact and row-majored.
|
||||||
|
*/
|
||||||
|
int64_t* strides;
|
||||||
|
/*! \brief The offset in bytes to the beginning pointer to data */
|
||||||
|
uint64_t byte_offset;
|
||||||
|
} DLTensor;
|
||||||
|
|
||||||
|
/*!
|
||||||
|
* \brief C Tensor object, manage memory of DLTensor. This data structure is
|
||||||
|
* intended to facilitate the borrowing of DLTensor by another framework. It is
|
||||||
|
* not meant to transfer the tensor. When the borrowing framework doesn't need
|
||||||
|
* the tensor, it should call the deleter to notify the host that the resource
|
||||||
|
* is no longer needed.
|
||||||
|
*/
|
||||||
|
typedef struct DLManagedTensor {
|
||||||
|
/*! \brief DLTensor which is being memory managed */
|
||||||
|
DLTensor dl_tensor;
|
||||||
|
/*! \brief the context of the original host framework of DLManagedTensor in
|
||||||
|
* which DLManagedTensor is used in the framework. It can also be NULL.
|
||||||
|
*/
|
||||||
|
void * manager_ctx;
|
||||||
|
/*! \brief Destructor signature void (*)(void*) - this should be called
|
||||||
|
* to destruct manager_ctx which holds the DLManagedTensor. It can be NULL
|
||||||
|
* if there is no way for the caller to provide a reasonable destructor.
|
||||||
|
* The destructors deletes the argument self as well.
|
||||||
|
*/
|
||||||
|
void (*deleter)(struct DLManagedTensor * self);
|
||||||
|
} DLManagedTensor;
|
||||||
|
#ifdef __cplusplus
|
||||||
|
} // DLPACK_EXTERN_C
|
||||||
|
#endif
|
||||||
|
#endif // DLPACK_DLPACK_H_
|
6
third_party/dlpack/tests/scripts/task_build.sh
vendored
Executable file
6
third_party/dlpack/tests/scripts/task_build.sh
vendored
Executable file
@@ -0,0 +1,6 @@
|
|||||||
|
#!/bin/bash
|
||||||
|
make || exit -1
|
||||||
|
mkdir -p build
|
||||||
|
cd build
|
||||||
|
cmake .. || exit -1
|
||||||
|
make || exit -1
|
23
third_party/dlpack/tests/scripts/task_lint.sh
vendored
Executable file
23
third_party/dlpack/tests/scripts/task_lint.sh
vendored
Executable file
@@ -0,0 +1,23 @@
|
|||||||
|
#!/bin/bash
|
||||||
|
mkdir -p bin
|
||||||
|
|
||||||
|
if [ ! -f bin/lint.py ]; then
|
||||||
|
echo "Grab linter ..."
|
||||||
|
wget https://raw.githubusercontent.com/dmlc/dmlc-core/main/scripts/lint.py
|
||||||
|
mv lint.py bin/lint.py
|
||||||
|
fi
|
||||||
|
|
||||||
|
echo "Check codestyle of c++ code..."
|
||||||
|
python bin/lint.py dlpack cpp include contrib
|
||||||
|
|
||||||
|
echo "Check doxygen generation..."
|
||||||
|
make doc 2>log.txt
|
||||||
|
(cat log.txt| grep -v ENABLE_PREPROCESSING |grep -v "unsupported tag") > logclean.txt
|
||||||
|
echo "---------Error Log----------"
|
||||||
|
cat logclean.txt
|
||||||
|
echo "----------------------------"
|
||||||
|
(cat logclean.txt|grep warning) && exit -1
|
||||||
|
(cat logclean.txt|grep error) && exit -1
|
||||||
|
rm logclean.txt
|
||||||
|
rm log.txt
|
||||||
|
echo "All checks passed..."
|
Reference in New Issue
Block a user