Compare commits
70 Commits
Author | SHA1 | Date | |
---|---|---|---|
887d5db635 | |||
ecbae3f79b | |||
b54979aa74 | |||
724f3d0bb9 | |||
06c5710412 | |||
b45fae576f | |||
55f093156d | |||
ff69c329cc | |||
d0accd6e0b | |||
89b7f045a4 | |||
46db4f3fe8 | |||
ef0bb7550d | |||
730b3dad83 | |||
9c69e94102 | |||
bd9135820a | |||
![]() |
4404590c44 | ||
441110afd6 | |||
a81e18eebc | |||
8f1046632c | |||
37ee82037e | |||
028deaa2ce | |||
ce4a106921 | |||
ef5aaac254 | |||
86eb20a33b | |||
ba275dd62a | |||
c2a28fc333 | |||
71cbe31915 | |||
f6fdb2228b | |||
b869686f6c | |||
3253ad2918 | |||
770c0167af | |||
c8bee63942 | |||
3e00543028 | |||
0e51c3302f | |||
c9837f5e68 | |||
a3f5042c64 | |||
53d607db2c | |||
e009ceb826 | |||
8c1f068406 | |||
a099849caa | |||
71a39e90e4 | |||
372d4a42f5 | |||
12963bae91 | |||
b71d86f735 | |||
2d654d2d2a | |||
e6315ee186 | |||
75e792b86c | |||
617ed541e3 | |||
![]() |
aa020eee0c | ||
![]() |
b110886935 | ||
![]() |
6bc6880d47 | ||
![]() |
b51b63dedc | ||
![]() |
601613b6c4 | ||
5e93ed0706 | |||
![]() |
1345f6e4c9 | ||
![]() |
1a381f01b7 | ||
![]() |
b42a170650 | ||
![]() |
1d30a5f237 | ||
![]() |
00a869ff07 | ||
![]() |
69f5e5496b | ||
![]() |
56f83b8132 | ||
![]() |
cda6673f85 | ||
![]() |
1d7a512fd4 | ||
658a73a284 | |||
71c9bb7a5c | |||
45efb06aa2 | |||
538b6f4659 | |||
![]() |
08e2589dfd | ||
![]() |
d098273816 | ||
![]() |
2f9c550323 |
106
.github/workflows/pypi-deploy.yml
vendored
Normal file
106
.github/workflows/pypi-deploy.yml
vendored
Normal file
@@ -0,0 +1,106 @@
|
||||
|
||||
name: Build Python Wheels
|
||||
|
||||
on:
|
||||
push:
|
||||
#branches: [master]
|
||||
tags: ["v[0-9]+.[0-9]+.[0-9]+"]
|
||||
|
||||
jobs:
|
||||
|
||||
build_setup:
|
||||
name: Prepare environment for wheel builds
|
||||
runs-on: ubuntu-20.04
|
||||
steps:
|
||||
- uses: actions/checkout@v2
|
||||
- name: Prepare wheel build
|
||||
run: make -C python/ setup
|
||||
- name: Store wheel configuration files
|
||||
uses: actions/upload-artifact@v2
|
||||
with:
|
||||
name: wheel-config
|
||||
path: python/
|
||||
|
||||
build_wheels:
|
||||
name: Build binary wheels on ${{ matrix.os }}
|
||||
runs-on: ${{ matrix.os }}
|
||||
needs: [build_setup]
|
||||
strategy:
|
||||
matrix:
|
||||
os: [ubuntu-20.04, windows-2019]
|
||||
arch: [auto32, auto64, aarch64]
|
||||
|
||||
steps:
|
||||
- uses: actions/checkout@v2
|
||||
- uses: actions/setup-python@v2
|
||||
- name: Install cibuildwheel
|
||||
run: python -m pip install cibuildwheel==2.1.2
|
||||
- name: Get wheel configuration files
|
||||
uses: actions/download-artifact@v2
|
||||
with:
|
||||
name: wheel-config
|
||||
path: python/
|
||||
- name: Get vcpkg repo
|
||||
if: matrix.os == 'windows-2019'
|
||||
run: git clone https://github.com/Microsoft/vcpkg.git
|
||||
- name: Build and install vcpkg and install libiconv for windows
|
||||
if: matrix.os == 'windows-2019'
|
||||
run: |
|
||||
.\vcpkg\bootstrap-vcpkg.bat
|
||||
vcpkg install libiconv
|
||||
working-directory: vcpkg/
|
||||
- name: Build wheels
|
||||
run: python -m cibuildwheel --output-dir wheelhouse
|
||||
working-directory: python/
|
||||
# env:
|
||||
# CIBW_BUILD: cp36-* cp37-* cp38-* cp39-* cp310-*
|
||||
- name: Store binary wheels
|
||||
uses: actions/upload-artifact@v2
|
||||
with:
|
||||
name: binary-wheels
|
||||
path: python/wheelhouse/*.whl
|
||||
|
||||
build_sdist:
|
||||
name: Build source distribution
|
||||
runs-on: ubuntu-latest
|
||||
needs: [build_setup]
|
||||
steps:
|
||||
- uses: actions/checkout@v2
|
||||
- name: Install cython
|
||||
run: python -m pip install cython==0.29.24
|
||||
- name: Get wheel configuration files
|
||||
uses: actions/download-artifact@v2
|
||||
with:
|
||||
name: wheel-config
|
||||
path: python/
|
||||
- name: Build sdist
|
||||
run: python setup.py sdist
|
||||
working-directory: python/
|
||||
- name: Store source wheels
|
||||
uses: actions/upload-artifact@v2
|
||||
with:
|
||||
name: source-wheels
|
||||
path: python/dist/*.tar.gz
|
||||
|
||||
|
||||
upload_pypi:
|
||||
name: Upload binary wheels to PyPI
|
||||
runs-on: ubuntu-20.04
|
||||
needs: [build_wheels, build_sdist]
|
||||
|
||||
steps:
|
||||
- name: Get source wheels
|
||||
uses: actions/download-artifact@v2
|
||||
with:
|
||||
name: source-wheels
|
||||
path: dist/
|
||||
- name: Get binary wheels
|
||||
uses: actions/download-artifact@v2
|
||||
with:
|
||||
name: binary-wheels
|
||||
path: dist/
|
||||
|
||||
- uses: pypa/gh-action-pypi-publish@release/v1
|
||||
with:
|
||||
user: __token__
|
||||
password: ${{ secrets.IMCTERMITE_GITHUB_WORKFLOW_PYPI_API_TOKEN }}
|
14
.gitignore
vendored
14
.gitignore
vendored
@@ -2,6 +2,7 @@
|
||||
eatraw
|
||||
eatdev
|
||||
imctermite
|
||||
IMCtermite
|
||||
|
||||
nohup.out
|
||||
|
||||
@@ -15,6 +16,7 @@ cython/*.cpp
|
||||
|
||||
*.log
|
||||
*.so
|
||||
*.pyd
|
||||
*.o
|
||||
*.csv
|
||||
*.parquet
|
||||
@@ -29,3 +31,15 @@ pip/README.md
|
||||
pip/LICENSE
|
||||
pip/*egg-info
|
||||
pip/dist/
|
||||
pip/build/
|
||||
|
||||
python/README.md
|
||||
python/README.rst
|
||||
python/LICENSE
|
||||
python/build
|
||||
python/*.egg-info
|
||||
python/dist
|
||||
python/*.soc
|
||||
python/lib/
|
||||
python/*.cpp
|
||||
python/wheelhouse/
|
||||
|
43
README.md
43
README.md
@@ -4,6 +4,7 @@
|
||||
[](https://lgtm.com/projects/g/RecordEvolution/IMCtermite/context:python)
|
||||
[](https://img.shields.io/github/license/RecordEvolution/IMCtermite)
|
||||
[](https://img.shields.io/github/stars/RecordEvolution/IMCtermite)
|
||||

|
||||
|
||||
# IMCtermite
|
||||
|
||||
@@ -30,6 +31,8 @@ On the [Record Evolution Platform](https://www.record-evolution.de/en/home-en/),
|
||||
|
||||
## File format
|
||||
|
||||
[Warning: Take a look at [this issue](https://github.com/RecordEvolution/IMCtermite/issues/14) when reading this section regarding the file format.]
|
||||
|
||||
A data file of the _IMC Bus Format_ type with the extension _.raw_ is a _mixed text/binary
|
||||
file_ featuring a set of markers (keys) that indicate the start of various blocks
|
||||
of data that provide meta information and the actual measurement data. Every single
|
||||
@@ -149,8 +152,11 @@ To install the latest version simply do
|
||||
python3 -m pip install IMCtermite
|
||||
```
|
||||
|
||||
Note, that _python3_setuptools_ and _gcc version >= 10.2.0_ are required to
|
||||
successfully install and use it.
|
||||
which provides binary wheels for multiple architectures on _Windows_ and _Linux_
|
||||
and most _Python 3.x_ distributions. However, if your platform/architecture is
|
||||
not supported you can still compile the source distribution yourself, which
|
||||
requires _python3_setuptools_ and an up-to-date compiler supporting C++11
|
||||
standard (e.g. _gcc version >= 10.2.0_).
|
||||
|
||||
## Usage
|
||||
|
||||
@@ -171,34 +177,38 @@ Options:
|
||||
-c, --listchannels list channels
|
||||
-b, --listblocks list IMC key-blocks
|
||||
-d, --output output directory to print channels
|
||||
-s, --delimiter csv delimiter/separator char for output
|
||||
-h, --help show this help message
|
||||
-v, --version display version
|
||||
```
|
||||
|
||||
For instance, to show a list of all channels included in `sample-data.raw`, you
|
||||
do `imctermite sample-data.raw --listchannels`. No output files are
|
||||
written by default. Output files are written only when an existing (!) directory is provided as argument to
|
||||
the `--output` option.
|
||||
written by default. Output files are written only when an existing (!) directory
|
||||
is provided as argument to the `--output` option. By default, every output file
|
||||
is written using a `,` delimiter. You may provide any custom separator with the
|
||||
option `--delimiter`. For example, in order to use `|`, the binary is called with
|
||||
options `imctermite sample-data.raw -b -c -s '|'`.
|
||||
|
||||
### Python
|
||||
|
||||
Given the `imctermite` module is available, we can import it and declare an instance
|
||||
Given the `IMCtermite` module is available, we can import it and declare an instance
|
||||
of it by passing a _raw_ file to the constructor:
|
||||
|
||||
```Python
|
||||
import imc_termite
|
||||
import IMCtermite
|
||||
|
||||
imcraw = imc_termite.imctermite(b"sample/sampleA.raw")
|
||||
imcraw = IMCtermite.imctermite(b"sample/sampleA.raw")
|
||||
```
|
||||
|
||||
An example of how to create an instance and obtain the list of channels is:
|
||||
|
||||
```Python
|
||||
import imc_termite
|
||||
import IMCtermite
|
||||
|
||||
# declare and initialize instance of "imctermite" by passing a raw-file
|
||||
try :
|
||||
imcraw = imc_termite.imctermite(b"samples/sampleA.raw")
|
||||
imcraw = IMCtermite.imctermite(b"samples/sampleA.raw")
|
||||
except RuntimeError as e :
|
||||
print("failed to load/parse raw-file: " + str(e))
|
||||
|
||||
@@ -207,9 +217,9 @@ channels = imcraw.get_channels(False)
|
||||
print(channels)
|
||||
```
|
||||
|
||||
A more complete [example](python/usage.py), including the methods for obtaining the
|
||||
channels, i.a. their data and/or directly printing them to files, can be found
|
||||
in the Python folder.
|
||||
A more complete [example](python/examples/usage.py), including the methods for
|
||||
obtaining the channels, i.a. their data and/or directly printing them to files,
|
||||
can be found in the `python/examples` folder.
|
||||
|
||||
## References
|
||||
|
||||
@@ -217,5 +227,12 @@ in the Python folder.
|
||||
- https://www.imc-tm.de/produkte/messtechnik-hardware/imc-cronosflex/ueberblick/
|
||||
- https://cython.readthedocs.io/en/latest/src/userguide/wrapping_CPlusPlus.html
|
||||
- https://github.com/Apollo3zehn/ImcFamosFile
|
||||
- https://apollo3zehn.github.io/ImcFamosFile/api/ImcFamosFile.FamosFileKeyType.html
|
||||
- https://pypi.org/help/#apitoken
|
||||
|
||||
- https://sgoel.dev/posts/uploading-binary-wheels-to-pypi-from-github-actions/
|
||||
- https://docs.github.com/en/actions/reference/workflow-syntax-for-github-actions#jobsjob_idstepsrun
|
||||
- https://github.com/pypa/cibuildwheel/blob/main/examples/github-deploy.yml
|
||||
- https://cibuildwheel.readthedocs.io/en/stable/deliver-to-pypi/
|
||||
- https://www.gnu.org/software/libiconv/
|
||||
- https://vcpkg.io/en/packages.html
|
||||
- https://vcpkg.io/en/getting-started
|
||||
|
@@ -1,111 +0,0 @@
|
||||
<?xml version="1.0" encoding="UTF-8" standalone="no"?>
|
||||
<svg
|
||||
xmlns:dc="http://purl.org/dc/elements/1.1/"
|
||||
xmlns:cc="http://creativecommons.org/ns#"
|
||||
xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#"
|
||||
xmlns:svg="http://www.w3.org/2000/svg"
|
||||
xmlns="http://www.w3.org/2000/svg"
|
||||
xmlns:sodipodi="http://sodipodi.sourceforge.net/DTD/sodipodi-0.dtd"
|
||||
xmlns:inkscape="http://www.inkscape.org/namespaces/inkscape"
|
||||
height="78.080002"
|
||||
width="325.72"
|
||||
id="svg3945"
|
||||
version="1.1"
|
||||
viewBox="0 0 325.72 78.08"
|
||||
sodipodi:docname="imctermite.svg"
|
||||
inkscape:version="1.0.1 (3bc2e813f5, 2020-09-07)">
|
||||
<sodipodi:namedview
|
||||
pagecolor="#ffffff"
|
||||
bordercolor="#666666"
|
||||
borderopacity="1"
|
||||
objecttolerance="10"
|
||||
gridtolerance="10"
|
||||
guidetolerance="10"
|
||||
inkscape:pageopacity="0"
|
||||
inkscape:pageshadow="2"
|
||||
inkscape:window-width="2048"
|
||||
inkscape:window-height="1088"
|
||||
id="namedview18"
|
||||
showgrid="false"
|
||||
inkscape:zoom="2.5577614"
|
||||
inkscape:cx="138.13369"
|
||||
inkscape:cy="-27.086877"
|
||||
inkscape:window-x="0"
|
||||
inkscape:window-y="32"
|
||||
inkscape:window-maximized="1"
|
||||
inkscape:current-layer="svg3945"
|
||||
inkscape:document-rotation="0" />
|
||||
<metadata
|
||||
id="metadata3951">
|
||||
<rdf:RDF>
|
||||
<cc:Work
|
||||
rdf:about="">
|
||||
<dc:format>image/svg+xml</dc:format>
|
||||
<dc:type
|
||||
rdf:resource="http://purl.org/dc/dcmitype/StillImage" />
|
||||
<dc:title>flasher</dc:title>
|
||||
</cc:Work>
|
||||
</rdf:RDF>
|
||||
</metadata>
|
||||
<defs
|
||||
id="defs3949" />
|
||||
<title
|
||||
id="title3916">flasher</title>
|
||||
<g
|
||||
id="logog"
|
||||
transform="translate(0,0.99981694)">
|
||||
<path
|
||||
style="fill:#364d5c"
|
||||
transform="translate(-2.04,-1.15)"
|
||||
d="m 32.86,2 -13,7.5 v 0 h -0.05 v 0 l -0.48,0.28 c -4.27,2.46 -5.68,11.38 -6.06,14.75 L 36.2,11.33 c 0.39,-0.19 7.6,-3.69 13.57,-3.69 h 0.14 L 40.13,2 a 8.15,8.15 0 0 0 -7.27,0"
|
||||
id="path138"
|
||||
inkscape:connector-curvature="0" />
|
||||
<path
|
||||
style="fill:#364d5c"
|
||||
transform="translate(-2.04,-1.15)"
|
||||
d="M 5.68,17.69 A 8.2,8.2 0 0 0 2,24 v 15.78 c 0,4.9 7,10.48 9.75,12.46 V 25.77 c 0,-0.44 0.6,-8.55 3.65,-13.72 z"
|
||||
id="path142"
|
||||
inkscape:connector-curvature="0" />
|
||||
<path
|
||||
style="fill:#364d5c"
|
||||
transform="translate(-2.04,-1.15)"
|
||||
d="m 12.1,54.12 v 0 C 11.74,53.88 5,49.41 2,44.24 v 11.14 a 8.2,8.2 0 0 0 3.64,6.3 l 13.5,7.79 c 4.28,2.46 12.7,-0.77 15.81,-2.12 z"
|
||||
id="path146"
|
||||
inkscape:connector-curvature="0" />
|
||||
<path
|
||||
style="fill:#364d5c"
|
||||
transform="translate(-2.04,-1.15)"
|
||||
d="m 36.79,68 c -0.4,0.19 -7.71,3.75 -13.71,3.69 l 9.78,5.64 a 8.15,8.15 0 0 0 7.27,0 l 13.51,-7.8 c 4.27,-2.46 5.68,-11.39 6.06,-14.75 z"
|
||||
id="path150"
|
||||
inkscape:connector-curvature="0" />
|
||||
<path
|
||||
style="fill:#364d5c"
|
||||
transform="translate(-2.04,-1.15)"
|
||||
d="M 61.2,27.13 V 53.6 c 0,0.44 -0.6,8.55 -3.65,13.72 l 9.77,-5.64 A 8.2,8.2 0 0 0 71,55.38 V 39.59 c 0,-4.94 -7,-10.5 -9.75,-12.46"
|
||||
id="path154"
|
||||
inkscape:connector-curvature="0" />
|
||||
<path
|
||||
style="fill:#364d5c"
|
||||
transform="translate(-2.04,-1.15)"
|
||||
d="M 67.31,17.69 53.81,9.9 C 49.53,7.44 41.11,10.67 38,12 l 22.85,13.23 v 0 a 43.43,43.43 0 0 1 5.7,4.51 24,24 0 0 1 4.45,5.35 V 24 a 8.2,8.2 0 0 0 -3.64,-6.3"
|
||||
id="path158"
|
||||
inkscape:connector-curvature="0" />
|
||||
</g>
|
||||
<g
|
||||
id="re"
|
||||
transform="translate(0,0.99981694)" />
|
||||
<text
|
||||
id="text3955"
|
||||
y="55.47554"
|
||||
x="74.101189"
|
||||
style="font-style:normal;font-weight:normal;font-size:40px;line-height:1.25;font-family:sans-serif;letter-spacing:0px;word-spacing:0px;fill:#364d5c;fill-opacity:1;stroke:none"
|
||||
xml:space="preserve"><tspan
|
||||
style="font-style:normal;font-variant:normal;font-weight:bold;font-stretch:normal;font-size:44px;font-family:sans-serif;-inkscape-font-specification:'sans-serif, Bold';font-variant-ligatures:normal;font-variant-caps:normal;font-variant-numeric:normal;font-feature-settings:normal;text-align:start;writing-mode:lr-tb;text-anchor:start;fill:#364d5c;fill-opacity:1"
|
||||
y="55.47554"
|
||||
x="74.101189"
|
||||
id="tspan3953"><tspan
|
||||
id="tspan24"
|
||||
style="font-style:normal;font-variant:normal;font-weight:bold;font-stretch:normal;font-family:sans-serif;-inkscape-font-specification:'sans-serif Bold'">IMC</tspan><tspan
|
||||
id="tspan3845"
|
||||
style="font-style:normal;font-variant:normal;font-weight:normal;font-stretch:normal;font-family:sans-serif;-inkscape-font-specification:sans-serif">termite</tspan> </tspan></text>
|
||||
</svg>
|
Before Width: | Height: | Size: 4.5 KiB |
@@ -1,112 +0,0 @@
|
||||
<?xml version="1.0" encoding="UTF-8" standalone="no"?>
|
||||
<svg
|
||||
xmlns:dc="http://purl.org/dc/elements/1.1/"
|
||||
xmlns:cc="http://creativecommons.org/ns#"
|
||||
xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#"
|
||||
xmlns:svg="http://www.w3.org/2000/svg"
|
||||
xmlns="http://www.w3.org/2000/svg"
|
||||
xmlns:sodipodi="http://sodipodi.sourceforge.net/DTD/sodipodi-0.dtd"
|
||||
xmlns:inkscape="http://www.inkscape.org/namespaces/inkscape"
|
||||
height="78.080002"
|
||||
width="290.72"
|
||||
id="svg3945"
|
||||
version="1.1"
|
||||
viewBox="0 0 290.72 78.08"
|
||||
sodipodi:docname="raweater.svg"
|
||||
inkscape:version="0.92.5 (2060ec1f9f, 2020-04-08)">
|
||||
<sodipodi:namedview
|
||||
pagecolor="#ffffff"
|
||||
bordercolor="#666666"
|
||||
borderopacity="1"
|
||||
objecttolerance="10"
|
||||
gridtolerance="10"
|
||||
guidetolerance="10"
|
||||
inkscape:pageopacity="0"
|
||||
inkscape:pageshadow="2"
|
||||
inkscape:window-width="1360"
|
||||
inkscape:window-height="704"
|
||||
id="namedview18"
|
||||
showgrid="false"
|
||||
inkscape:zoom="0.90430522"
|
||||
inkscape:cx="191.86"
|
||||
inkscape:cy="38.540001"
|
||||
inkscape:window-x="0"
|
||||
inkscape:window-y="27"
|
||||
inkscape:window-maximized="1"
|
||||
inkscape:current-layer="svg3945" />
|
||||
<metadata
|
||||
id="metadata3951">
|
||||
<rdf:RDF>
|
||||
<cc:Work
|
||||
rdf:about="">
|
||||
<dc:format>image/svg+xml</dc:format>
|
||||
<dc:type
|
||||
rdf:resource="http://purl.org/dc/dcmitype/StillImage" />
|
||||
<dc:title>flasher</dc:title>
|
||||
</cc:Work>
|
||||
</rdf:RDF>
|
||||
</metadata>
|
||||
<defs
|
||||
id="defs3949" />
|
||||
<title
|
||||
id="title3916">flasher</title>
|
||||
<g
|
||||
id="logog"
|
||||
transform="translate(0,0.99981694)">
|
||||
<path
|
||||
style="fill:#364d5c"
|
||||
transform="translate(-2.04,-1.15)"
|
||||
d="m 32.86,2 -13,7.5 v 0 h -0.05 v 0 l -0.48,0.28 c -4.27,2.46 -5.68,11.38 -6.06,14.75 L 36.2,11.33 c 0.39,-0.19 7.6,-3.69 13.57,-3.69 h 0.14 L 40.13,2 a 8.15,8.15 0 0 0 -7.27,0"
|
||||
id="path138"
|
||||
inkscape:connector-curvature="0" />
|
||||
<path
|
||||
style="fill:#364d5c"
|
||||
transform="translate(-2.04,-1.15)"
|
||||
d="M 5.68,17.69 A 8.2,8.2 0 0 0 2,24 v 15.78 c 0,4.9 7,10.48 9.75,12.46 V 25.77 c 0,-0.44 0.6,-8.55 3.65,-13.72 z"
|
||||
id="path142"
|
||||
inkscape:connector-curvature="0" />
|
||||
<path
|
||||
style="fill:#364d5c"
|
||||
transform="translate(-2.04,-1.15)"
|
||||
d="m 12.1,54.12 v 0 C 11.74,53.88 5,49.41 2,44.24 v 11.14 a 8.2,8.2 0 0 0 3.64,6.3 l 13.5,7.79 c 4.28,2.46 12.7,-0.77 15.81,-2.12 z"
|
||||
id="path146"
|
||||
inkscape:connector-curvature="0" />
|
||||
<path
|
||||
style="fill:#364d5c"
|
||||
transform="translate(-2.04,-1.15)"
|
||||
d="m 36.79,68 c -0.4,0.19 -7.71,3.75 -13.71,3.69 l 9.78,5.64 a 8.15,8.15 0 0 0 7.27,0 l 13.51,-7.8 c 4.27,-2.46 5.68,-11.39 6.06,-14.75 z"
|
||||
id="path150"
|
||||
inkscape:connector-curvature="0" />
|
||||
<path
|
||||
style="fill:#364d5c"
|
||||
transform="translate(-2.04,-1.15)"
|
||||
d="M 61.2,27.13 V 53.6 c 0,0.44 -0.6,8.55 -3.65,13.72 l 9.77,-5.64 A 8.2,8.2 0 0 0 71,55.38 V 39.59 c 0,-4.94 -7,-10.5 -9.75,-12.46"
|
||||
id="path154"
|
||||
inkscape:connector-curvature="0" />
|
||||
<path
|
||||
style="fill:#364d5c"
|
||||
transform="translate(-2.04,-1.15)"
|
||||
d="M 67.31,17.69 53.81,9.9 C 49.53,7.44 41.11,10.67 38,12 l 22.85,13.23 v 0 a 43.43,43.43 0 0 1 5.7,4.51 24,24 0 0 1 4.45,5.35 V 24 a 8.2,8.2 0 0 0 -3.64,-6.3"
|
||||
id="path158"
|
||||
inkscape:connector-curvature="0" />
|
||||
</g>
|
||||
<g
|
||||
id="re"
|
||||
transform="translate(0,0.99981694)" />
|
||||
<text
|
||||
id="text3955"
|
||||
y="55.47554"
|
||||
x="74.101189"
|
||||
style="font-style:normal;font-weight:normal;font-size:40px;line-height:1.25;font-family:sans-serif;letter-spacing:0px;word-spacing:0px;fill:#364d5c;fill-opacity:1;stroke:none"
|
||||
xml:space="preserve"><tspan
|
||||
style="font-style:normal;font-variant:normal;font-weight:bold;font-stretch:normal;font-size:44px;font-family:sans-serif;-inkscape-font-specification:'sans-serif, Bold';font-variant-ligatures:normal;font-variant-caps:normal;font-variant-numeric:normal;font-feature-settings:normal;text-align:start;writing-mode:lr-tb;text-anchor:start;fill:#364d5c;fill-opacity:1"
|
||||
y="55.47554"
|
||||
x="74.101189"
|
||||
id="tspan3953"><tspan
|
||||
id="tspan24"
|
||||
style="font-style:normal;font-variant:normal;font-weight:bold;font-stretch:normal;font-family:sans-serif;-inkscape-font-specification:'sans-serif Bold'">R</tspan><tspan
|
||||
id="tspan3845"
|
||||
style="font-style:normal;font-variant:normal;font-weight:normal;font-stretch:normal;font-family:sans-serif;-inkscape-font-specification:sans-serif">aw<tspan
|
||||
style="font-style:normal;font-variant:normal;font-weight:bold;font-stretch:normal;font-family:sans-serif;-inkscape-font-specification:'sans-serif Bold'"
|
||||
id="tspan22">E</tspan>ater</tspan> </tspan></text>
|
||||
</svg>
|
Before Width: | Height: | Size: 4.6 KiB |
@@ -1,41 +0,0 @@
|
||||
# cython: language_level = 3
|
||||
# distutils: language = c++
|
||||
|
||||
# use some C++ STL libraries
|
||||
from libcpp.string cimport string
|
||||
from libcpp.vector cimport vector
|
||||
from libcpp cimport bool
|
||||
|
||||
# to include implemenation/definition file
|
||||
#cdef extern from "raweat.cpp":
|
||||
# pass
|
||||
|
||||
# these method names have to match the C definitions of the methods!!
|
||||
#
|
||||
# for how to overload the constructor see
|
||||
# https://cython.readthedocs.io/en/latest/src/userguide/wrapping_CPlusPlus.html
|
||||
# and propagating exceptions from C++ to Python
|
||||
# http://docs.cython.org/en/latest/src/userguide/wrapping_CPlusPlus.html#exceptions
|
||||
|
||||
cdef extern from "../lib/raweat.hpp":
|
||||
cdef cppclass raw_eater:
|
||||
# constructor(s)
|
||||
raw_eater() except +
|
||||
raw_eater(string) except +
|
||||
# set new file for decoding
|
||||
void set_file(string)
|
||||
# perform conversion (pass any C++ exceptions to Python)
|
||||
void setup_and_conversion() except +
|
||||
# get validity of data format
|
||||
bool get_valid()
|
||||
# get channel name and unit
|
||||
string get_name()
|
||||
string get_unit()
|
||||
# get time step and time unit
|
||||
double get_dt()
|
||||
string get_temp_unit()
|
||||
# get data array of time and measured quantity's channel
|
||||
vector[double] get_time()
|
||||
vector[double] get_data()
|
||||
# dump all data to .csv
|
||||
void write_table(const char*,char delimiter)
|
@@ -1,58 +0,0 @@
|
||||
|
||||
from raw_eater cimport raweater
|
||||
import numpy as np
|
||||
import re
|
||||
import os
|
||||
|
||||
cdef class raweater:
|
||||
|
||||
# C++ instance of class => stack allocated (requires nullary constructor!)
|
||||
cdef raw_eater rawit
|
||||
|
||||
# pointer to C++ instance (if there's no nullary constructor)
|
||||
# cdef raw_eater *rawit
|
||||
|
||||
def __cinit__(self, string rawfile = b''):
|
||||
if rawfile.decode() == "":
|
||||
self.rawit = raw_eater()
|
||||
# self.rawit = new raw_eater()
|
||||
else:
|
||||
if not os.path.isfile(rawfile) :
|
||||
raise ValueError("'" + str(rawfile) + "' does not exist")
|
||||
self.rawit = raw_eater(rawfile)
|
||||
# self.rawit = new raw_eater(rawfile)
|
||||
|
||||
# def __dealloc__(self):
|
||||
# del self.rawit
|
||||
|
||||
def set_file(self, string rawfile):
|
||||
if not os.path.isfile(rawfile) :
|
||||
raise ValueError("'" + str(rawfile) + "' does not exist")
|
||||
self.rawit.set_file(rawfile)
|
||||
|
||||
def do_conversion(self):
|
||||
self.rawit.setup_and_conversion()
|
||||
|
||||
def validity(self):
|
||||
return self.rawit.get_valid()
|
||||
|
||||
def channel_name(self):
|
||||
return self.rawit.get_name()
|
||||
|
||||
def unit(self):
|
||||
return self.rawit.get_unit()
|
||||
|
||||
def dt(self):
|
||||
return self.rawit.get_dt()
|
||||
|
||||
def time_unit(self):
|
||||
return self.rawit.get_temp_unit()
|
||||
|
||||
def get_time(self):
|
||||
return self.rawit.get_time()
|
||||
|
||||
def get_channel(self):
|
||||
return self.rawit.get_data()
|
||||
|
||||
def write_table(self, const char* csvfile, char delimiter):
|
||||
self.rawit.write_table(csvfile,delimiter)
|
@@ -1,37 +0,0 @@
|
||||
# cython: language_level = 3
|
||||
# distutils: language = c++
|
||||
|
||||
# use some C++ STL libraries
|
||||
from libcpp.string cimport string
|
||||
from libcpp.vector cimport vector
|
||||
from libcpp cimport bool
|
||||
|
||||
# these method names have to match the C++ definitions of the methods!!
|
||||
cdef extern from "../lib/rawmerge.hpp":
|
||||
cdef cppclass raw_merger:
|
||||
raw_merger(string) except +
|
||||
# get validity of data format
|
||||
bool get_valid()
|
||||
# get channel name and unit
|
||||
string get_name()
|
||||
string get_unit()
|
||||
# get time step and time unit
|
||||
double get_dt()
|
||||
string get_temp_unit()
|
||||
# get data array of time and measured quantity's channel
|
||||
vector[double] get_time()
|
||||
vector[double] get_data()
|
||||
# dump all data to .csv
|
||||
void write_table(const char*,char)
|
||||
# add channel and try to merge it (pass C++ exceptions to Python)
|
||||
bool add_channel(string) except +
|
||||
# get total number of (added) channels
|
||||
int get_num_channels()
|
||||
# get list of channel names
|
||||
vector[string] get_channel_names()
|
||||
# get data of particular channel
|
||||
vector[double] get_channel(int)
|
||||
# get total merged time series
|
||||
vector[double] get_time_series()
|
||||
# dump all channels to .csv
|
||||
void write_table_all(const char*,char)
|
@@ -1,58 +0,0 @@
|
||||
|
||||
# from <raw_meat> has to match name of .pxd file and cimport name of class defined in .pxd
|
||||
from raw_meat cimport raw_merger
|
||||
import numpy as np
|
||||
import re
|
||||
|
||||
cdef class rawmerger:
|
||||
|
||||
# pointer to C++ instance (since there's no nullary constructor)
|
||||
cdef raw_merger *rawit
|
||||
|
||||
def __cinit__(self, string rawfile):
|
||||
self.rawit = new raw_merger(rawfile)
|
||||
|
||||
def __dealloc__(self):
|
||||
del self.rawit
|
||||
|
||||
def validity(self):
|
||||
return self.rawit.get_valid()
|
||||
|
||||
def channel_name(self):
|
||||
return self.rawit.get_name()
|
||||
|
||||
def unit(self):
|
||||
return self.rawit.get_unit()
|
||||
|
||||
def dt(self):
|
||||
return self.rawit.get_dt()
|
||||
|
||||
def time_unit(self):
|
||||
return self.rawit.get_temp_unit()
|
||||
|
||||
def get_time(self):
|
||||
return self.rawit.get_time()
|
||||
|
||||
def get_channel(self):
|
||||
return self.rawit.get_data()
|
||||
|
||||
def write_table(self, const char* csvfile, char delimiter):
|
||||
return self.rawit.write_table(csvfile,delimiter)
|
||||
|
||||
def add_channel(self, string rawfile):
|
||||
return self.rawit.add_channel(rawfile)
|
||||
|
||||
def get_num_channels(self):
|
||||
return self.rawit.get_num_channels()
|
||||
|
||||
def get_channel_names(self):
|
||||
return self.rawit.get_channel_names()
|
||||
|
||||
def get_channel_by_index(self, int chidx):
|
||||
return self.rawit.get_channel(chidx)
|
||||
|
||||
def get_time_series(self):
|
||||
return self.rawit.get_time_series()
|
||||
|
||||
def write_table_all(self, const char* csvfile, char delimiter):
|
||||
return self.rawit.write_table_all(csvfile,delimiter)
|
@@ -1,24 +0,0 @@
|
||||
from distutils.core import setup
|
||||
from distutils.extension import Extension
|
||||
from Cython.Build import cythonize
|
||||
|
||||
extensions = Extension(
|
||||
name="imc_termite",
|
||||
sources=["cython/py_imc_termite.pyx"],
|
||||
# libraries=[""],
|
||||
# library_dirs=["lib"],
|
||||
include_dirs=["lib"],
|
||||
language='c++',
|
||||
extra_compile_args=['-std=c++17','-Wno-unused-variable'],
|
||||
extra_link_args=['-std=c++17'],
|
||||
)
|
||||
|
||||
setup(
|
||||
name="imc_termite",
|
||||
version='1.2.4',
|
||||
description='IMCtermite cython extension',
|
||||
author='Record Evolution GmbH',
|
||||
author_email='mario.fink@record-evolution.de',
|
||||
url='https://github.com/RecordEvolution/IMCtermite.git',
|
||||
ext_modules=cythonize(extensions,force=True)
|
||||
)
|
@@ -1,20 +0,0 @@
|
||||
from distutils.core import setup
|
||||
from distutils.extension import Extension
|
||||
from Cython.Build import cythonize
|
||||
|
||||
extensions = Extension(
|
||||
name="raw_eater",
|
||||
sources=["cython/raw_eater.pyx"],
|
||||
# libraries=[""],
|
||||
library_dirs=["src"],
|
||||
include_dirs=["src"],
|
||||
language='c++',
|
||||
extra_compile_args=['-std=c++11','-Wno-unused-variable'],
|
||||
extra_link_args=['-std=c++11'],
|
||||
#extra_objects=["lib/parquet/libarrow.so.200.0.0"],
|
||||
)
|
||||
|
||||
setup(
|
||||
name="raw_eater",
|
||||
ext_modules=cythonize(extensions)
|
||||
)
|
@@ -1,20 +0,0 @@
|
||||
from distutils.core import setup
|
||||
from distutils.extension import Extension
|
||||
from Cython.Build import cythonize
|
||||
|
||||
extensions = Extension(
|
||||
name="raw_meat",
|
||||
sources=["cython/raw_meat.pyx"],
|
||||
# libraries=[""],
|
||||
library_dirs=["src"],
|
||||
include_dirs=["src"],
|
||||
language='c++',
|
||||
extra_compile_args=['-std=c++11','-Wno-unused-variable'],
|
||||
extra_link_args=['-std=c++11'],
|
||||
#extra_objects=["lib/parquet/libarrow.so.200.0.0"],
|
||||
)
|
||||
|
||||
setup(
|
||||
name="raw_meat",
|
||||
ext_modules=cythonize(extensions)
|
||||
)
|
0
lib/half.hpp → lib/3rdparty/half.hpp
vendored
0
lib/half.hpp → lib/3rdparty/half.hpp
vendored
@@ -63,7 +63,7 @@ namespace imc
|
||||
std::cout<<"WARNING: invalid length parameter in "<<thekey_.name_<<"-block "
|
||||
<<"(block-end:"<<end_<<",buffer-size:"<<buffer_->size()<<")"
|
||||
<<" => resetting block-end to buffer-size\n";
|
||||
end_ = buffer_->size();
|
||||
end_ = (unsigned long int)(buffer_->size());
|
||||
}
|
||||
|
||||
try {
|
||||
|
@@ -3,9 +3,14 @@
|
||||
#ifndef IMCCHANNEL
|
||||
#define IMCCHANNEL
|
||||
|
||||
#include <sstream>
|
||||
#include "imc_datatype.hpp"
|
||||
#include "imc_conversion.hpp"
|
||||
#include <sstream>
|
||||
#include <math.h>
|
||||
#include <chrono>
|
||||
#include <ctime>
|
||||
#include <time.h>
|
||||
#include <iconv.h>
|
||||
|
||||
//---------------------------------------------------------------------------//
|
||||
|
||||
@@ -20,7 +25,7 @@ namespace imc
|
||||
// collect affiliate blocks for a single channel
|
||||
std::string CBuuid_, CGuuid_, CCuuid_, CNuuid_;
|
||||
std::string CDuuid_, CTuuid_, Cbuuid_, CPuuid_, CRuuid_, CSuuid_;
|
||||
std::string NTuuid_, NOuuid_;
|
||||
std::string NTuuid_, NOuuid_, NLuuid_;
|
||||
|
||||
// reset all members
|
||||
void reset()
|
||||
@@ -38,6 +43,7 @@ namespace imc
|
||||
CSuuid_.clear();
|
||||
NTuuid_.clear();
|
||||
NOuuid_.clear();
|
||||
NLuuid_.clear();
|
||||
}
|
||||
|
||||
// get info
|
||||
@@ -58,7 +64,8 @@ namespace imc
|
||||
<<std::setw(width)<<std::left<<"CSuuid:"<<CSuuid_<<"\n"
|
||||
//
|
||||
<<std::setw(width)<<std::left<<"NTuuid:"<<NTuuid_<<"\n"
|
||||
<<std::setw(width)<<std::left<<"NOuuid:"<<NOuuid_<<"\n";
|
||||
<<std::setw(width)<<std::left<<"NOuuid:"<<NOuuid_<<"\n"
|
||||
<<std::setw(width)<<std::left<<"NLuuid:"<<NLuuid_<<"\n";
|
||||
return ss.str();
|
||||
}
|
||||
|
||||
@@ -79,32 +86,56 @@ namespace imc
|
||||
<<"\",\"CSuuid\":\""<<CSuuid_
|
||||
<<"\",\"NTuuid\":\""<<NTuuid_
|
||||
<<"\",\"NOuuid\":\""<<NOuuid_
|
||||
<<"\",\"NLuuid\":\""<<NLuuid_
|
||||
<<"\"}";
|
||||
return ss.str();
|
||||
}
|
||||
};
|
||||
|
||||
// adjust stream object
|
||||
void customize_stream(std::ostream& stout, int prec, bool fixed)
|
||||
{
|
||||
if ( fixed )
|
||||
{
|
||||
stout<<std::setprecision(prec)<<std::fixed;
|
||||
}
|
||||
else
|
||||
{
|
||||
stout<<std::setprecision(prec);
|
||||
}
|
||||
}
|
||||
|
||||
// given a list of numeric objects, join it into a string
|
||||
template<typename dt>
|
||||
std::string joinvec(std::vector<dt> myvec, unsigned long int limit = 10)
|
||||
std::string joinvec(std::vector<dt> myvec, unsigned long int limit = 10, int prec = 10, bool fixed = true)
|
||||
{
|
||||
// include entire list for limit = - 1
|
||||
limit = (limit == 0) ? myvec.size() : limit;
|
||||
// include entire list for limit = 0
|
||||
unsigned long int myvecsize = (unsigned long int)myvec.size();
|
||||
limit = (limit == 0) ? myvecsize : limit;
|
||||
|
||||
std::stringstream ss;
|
||||
ss<<"[";
|
||||
if ( myvec.size() <= limit )
|
||||
{
|
||||
for ( dt el: myvec ) ss<<std::setprecision(10)<<el<<",";
|
||||
for ( dt el: myvec )
|
||||
{
|
||||
customize_stream(ss,prec,fixed);
|
||||
ss<<el<<",";
|
||||
}
|
||||
}
|
||||
else
|
||||
{
|
||||
unsigned long int heals = (unsigned long int)(limit/2.);
|
||||
for ( unsigned long int i = 0; i < heals; i++ ) ss<<myvec[i]<<",";
|
||||
ss<<"...";
|
||||
for ( unsigned long int i = myvec.size()-heals; i < myvec.size(); i++ )
|
||||
unsigned long int heals = limit/2;
|
||||
for ( unsigned long int i = 0; i < heals; i++ )
|
||||
{
|
||||
ss<<std::setprecision(10)<<myvec[i]<<",";
|
||||
customize_stream(ss,prec,fixed);
|
||||
ss<<myvec[i]<<",";
|
||||
}
|
||||
ss<<"...";
|
||||
for ( unsigned long int i = myvecsize-heals; i < myvecsize; i++ )
|
||||
{
|
||||
customize_stream(ss,prec,fixed);
|
||||
ss<<myvec[i]<<",";
|
||||
}
|
||||
}
|
||||
std::string sumstr = ss.str();
|
||||
@@ -113,6 +144,78 @@ namespace imc
|
||||
return sumstr;
|
||||
}
|
||||
|
||||
// convert encoding of any descriptions, channel-names, units etc.
|
||||
class iconverter
|
||||
{
|
||||
std::string in_enc_, out_enc_;
|
||||
iconv_t cd_;
|
||||
size_t out_buffer_size_;
|
||||
|
||||
public:
|
||||
|
||||
iconverter(std::string in_enc, std::string out_enc, size_t out_buffer_size = 1024) :
|
||||
in_enc_(in_enc), out_enc_(out_enc), out_buffer_size_(out_buffer_size)
|
||||
{
|
||||
// allocate descriptor for character set conversion
|
||||
// (https://man7.org/linux/man-pages/man3/iconv_open.3.html)
|
||||
cd_ = iconv_open(out_enc.c_str(), in_enc.c_str());
|
||||
|
||||
if ( (iconv_t)-1 == cd_ )
|
||||
{
|
||||
if ( errno == EINVAL )
|
||||
{
|
||||
std::string errmsg = std::string("The encoding conversion from ") + in_enc
|
||||
+ std::string(" to ") + out_enc + std::string(" is not supported by the implementation.");
|
||||
throw std::runtime_error(errmsg);
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
void convert(std::string &astring)
|
||||
{
|
||||
if ( astring.empty() ) return;
|
||||
|
||||
std::vector<char> in_buffer(astring.begin(),astring.end());
|
||||
char *inbuf = &in_buffer[0];
|
||||
size_t inbytes = in_buffer.size();
|
||||
|
||||
std::vector<char> out_buffer(out_buffer_size_);
|
||||
char *outbuf = &out_buffer[0];
|
||||
size_t outbytes = out_buffer.size();
|
||||
|
||||
// perform character set conversion
|
||||
// ( - https://man7.org/linux/man-pages/man3/iconv.3.html
|
||||
// - https://www.ibm.com/docs/en/zos/2.2.0?topic=functions-iconv-code-conversion )
|
||||
while ( inbytes > 0 )
|
||||
{
|
||||
size_t res = iconv(cd_,&inbuf,&inbytes,&outbuf,&outbytes);
|
||||
|
||||
if ( (size_t)-1 == res )
|
||||
{
|
||||
std::string errmsg;
|
||||
if ( errno == EILSEQ )
|
||||
{
|
||||
errmsg = std::string("An invalid multibyte sequence is encountered in the input.");
|
||||
throw std::runtime_error(errmsg);
|
||||
}
|
||||
else if ( errno == EINVAL )
|
||||
{
|
||||
errmsg = std::string("An incomplete multibyte sequence is encountered in the input")
|
||||
+ std::string(" and the input byte sequence terminates after it.");
|
||||
}
|
||||
else if ( errno == E2BIG )
|
||||
{
|
||||
errmsg = std::string("The output buffer has no more room for the next converted character.");
|
||||
}
|
||||
throw std::runtime_error(errmsg);
|
||||
}
|
||||
}
|
||||
|
||||
std::string outstring(out_buffer.begin(),out_buffer.end()-outbytes);
|
||||
astring = outstring;
|
||||
}
|
||||
};
|
||||
|
||||
// channel
|
||||
struct channel
|
||||
{
|
||||
@@ -122,18 +225,23 @@ namespace imc
|
||||
std::vector<unsigned char>* buffer_;
|
||||
|
||||
// collect meta-data of channels according to env,
|
||||
// just everything valueable in here
|
||||
// just everything valueable in here
|
||||
std::string uuid_;
|
||||
std::string name_, comment_;
|
||||
std::string origin_, origin_comment_, text_;
|
||||
std::chrono::system_clock::time_point trigger_time_, absolute_trigger_time_;
|
||||
double trigger_time_frac_secs_;
|
||||
std::string language_code_, codepage_;
|
||||
std::string yname_, yunit_;
|
||||
std::string xname_, xunit_;
|
||||
double xstepwidth_, xoffset_;
|
||||
int xprec_;
|
||||
|
||||
// buffer and data
|
||||
int signbits_, num_bytes_;
|
||||
// unsigned long int byte_offset_;
|
||||
unsigned long int buffer_offset_, buffer_size_;
|
||||
long int addtime_;
|
||||
int datatp_;
|
||||
imc::datatype dattyp_;
|
||||
std::vector<imc::datatype> ydata_;
|
||||
@@ -185,6 +293,9 @@ namespace imc
|
||||
xunit_ = blocks_->at(chnenv_.CDuuid_).get_parameter(prms[5]);
|
||||
// TODO
|
||||
// xname_ = std::string("time");
|
||||
|
||||
// find appropriate precision for "xdata_" by means of "xstepwidth_"
|
||||
xprec_ = (xstepwidth_ > 0 ) ? (int)ceil(fabs(log10(xstepwidth_))) : 10;
|
||||
}
|
||||
|
||||
// extract associated CP data
|
||||
@@ -204,6 +315,7 @@ namespace imc
|
||||
buffer_offset_ = std::stoul(blocks_->at(chnenv_.Cbuuid_).get_parameter(prms[6]));
|
||||
buffer_size_ = std::stoul(blocks_->at(chnenv_.Cbuuid_).get_parameter(prms[7]));
|
||||
xoffset_ = std::stod(blocks_->at(chnenv_.Cbuuid_).get_parameter(prms[11]));
|
||||
addtime_ = (long int)std::stod(blocks_->at(chnenv_.Cbuuid_).get_parameter(prms[12]));
|
||||
}
|
||||
|
||||
// extract associated CR data
|
||||
@@ -233,26 +345,72 @@ namespace imc
|
||||
origin_comment_ = blocks_->at(chnenv_.NOuuid_).get_parameter(prms[6]);
|
||||
}
|
||||
|
||||
// extract associated NL data
|
||||
// codepage:
|
||||
// - http://www.iana.org/assignments/character-sets/character-sets.xhtml
|
||||
// - https://de.wikipedia.org/wiki/Zeichensatztabelle
|
||||
// language-code:
|
||||
// - https://docs.microsoft.com/en-us/openspecs/windows_protocols/ms-lcid/a9eac961-e77d-41a6-90a5-ce1a8b0cdb9c?redirectedfrom=MSDN
|
||||
if ( blocks_->count(chnenv_.NLuuid_) == 1 )
|
||||
{
|
||||
prms = blocks_->at(chnenv_.NLuuid_).get_parameters();
|
||||
codepage_ = blocks_->at(chnenv_.NLuuid_).get_parameter(prms[2]);
|
||||
language_code_ = blocks_->at(chnenv_.NLuuid_).get_parameter(prms[3]);
|
||||
}
|
||||
|
||||
// obtain NT data
|
||||
// - https://en.cppreference.com/w/cpp/chrono/c/tm
|
||||
// - https://en.cppreference.com/w/cpp/io/manip/put_time
|
||||
if ( blocks_->count(chnenv_.NTuuid_) == 1 )
|
||||
{
|
||||
prms = blocks_->at(chnenv_.NTuuid_).get_parameters();
|
||||
//std::tm tm{};
|
||||
std::tm tms = std::tm();
|
||||
tms.tm_mday = std::stoi(blocks_->at(chnenv_.NTuuid_).get_parameter(prms[2]));
|
||||
tms.tm_mon = std::stoi(blocks_->at(chnenv_.NTuuid_).get_parameter(prms[3])) - 1;
|
||||
tms.tm_year = std::stoi(blocks_->at(chnenv_.NTuuid_).get_parameter(prms[4])) - 1900;
|
||||
tms.tm_hour = std::stoi(blocks_->at(chnenv_.NTuuid_).get_parameter(prms[5]));
|
||||
tms.tm_min = std::stoi(blocks_->at(chnenv_.NTuuid_).get_parameter(prms[6]));
|
||||
long double secs = std::stold(blocks_->at(chnenv_.NTuuid_).get_parameter(prms[7]));
|
||||
double secs_int;
|
||||
trigger_time_frac_secs_ = modf((double)secs,&secs_int);
|
||||
tms.tm_sec = (int)secs_int;
|
||||
//tms.tm_isdst = -1;
|
||||
|
||||
// generate std::chrono::system_clock::time_point type
|
||||
// ( - https://www.gnu.org/software/libc/manual/html_node/Broken_002ddown-Time.html
|
||||
// - https://man7.org/linux/man-pages/man3/tzset.3.html )
|
||||
std::time_t ts = timegm(&tms); //std::mktime(&tms);
|
||||
trigger_time_ = std::chrono::system_clock::from_time_t(ts);
|
||||
}
|
||||
|
||||
// start converting binary buffer to imc::datatype
|
||||
if ( !chnenv_.CSuuid_.empty() ) convert_buffer();
|
||||
|
||||
// calculate absolute trigger-time
|
||||
absolute_trigger_time_ = trigger_time_ + std::chrono::seconds(addtime_);
|
||||
// + std::chrono::nanoseconds((long int)(trigger_time_frac_secs_*1.e9));
|
||||
|
||||
// convert any non-UTF-8 codepage to UTF-8
|
||||
convert_encoding();
|
||||
}
|
||||
|
||||
// convert buffer to actual datatype
|
||||
void convert_buffer()
|
||||
{
|
||||
// TODO no clue how/if/when to handle buffer offset/mask/subsequent_bytes
|
||||
// etc. and whatever that shit is!
|
||||
std::vector<imc::parameter> prms = blocks_->at(chnenv_.CSuuid_).get_parameters();
|
||||
if ( prms.size() < 4)
|
||||
{
|
||||
throw std::runtime_error("CS block is invalid and features to few parameters");
|
||||
}
|
||||
|
||||
// extract (channel dependent) part of buffer
|
||||
unsigned long int buffstrt = prms[3].begin();
|
||||
std::vector<unsigned char> CSbuffer( buffer_->begin()+buffstrt+1,
|
||||
buffer_->begin()+buffstrt+buffer_size_+1 );
|
||||
std::vector<unsigned char> CSbuffer( buffer_->begin()+buffstrt+buffer_offset_+1,
|
||||
buffer_->begin()+buffstrt+buffer_offset_+buffer_size_+1 );
|
||||
|
||||
// determine number of values in buffer
|
||||
unsigned long int num_values = CSbuffer.size()/(signbits_/8);
|
||||
unsigned long int num_values = (unsigned long int)(CSbuffer.size()/(signbits_/8));
|
||||
if ( num_values*(signbits_/8) != CSbuffer.size() )
|
||||
{
|
||||
throw std::runtime_error("CSbuffer and significant bits of datatype don't match");
|
||||
@@ -307,7 +465,7 @@ namespace imc
|
||||
// fill xdata_
|
||||
for ( unsigned long int i = 0; i < num_values; i++ )
|
||||
{
|
||||
xdata_.push_back(xoffset_+i*xstepwidth_);
|
||||
xdata_.push_back(xoffset_+(double)i*xstepwidth_);
|
||||
}
|
||||
|
||||
// employ data transformation
|
||||
@@ -315,27 +473,64 @@ namespace imc
|
||||
{
|
||||
for ( imc::datatype& el: ydata_ )
|
||||
{
|
||||
// std::cout<<"value:"<<el.as_double()<<"\n";
|
||||
el = imc::datatype(el.as_double()*factor_ + offset_);
|
||||
//std::cout<<"value:"<<el.as_double()<<"\n";
|
||||
double fact = ( factor_ == 0.0 ) ? 1.0 : factor_;
|
||||
el = imc::datatype(el.as_double()*fact + offset_);
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
// convert any description, units etc. to UTF-8 (by default)
|
||||
void convert_encoding()
|
||||
{
|
||||
if ( !codepage_.empty() )
|
||||
{
|
||||
// construct iconv-compatible name for respective codepage
|
||||
std::string cpn = std::string("CP") + codepage_;
|
||||
|
||||
// set up converter
|
||||
std::string utf = std::string("UTF-8");
|
||||
iconverter conv(cpn,utf);
|
||||
|
||||
conv.convert(name_);
|
||||
conv.convert(comment_);
|
||||
conv.convert(origin_);
|
||||
conv.convert(origin_comment_);
|
||||
conv.convert(text_);
|
||||
conv.convert(language_code_);
|
||||
conv.convert(yname_);
|
||||
conv.convert(yunit_);
|
||||
conv.convert(xname_);
|
||||
conv.convert(xunit_);
|
||||
conv.convert(group_name_);
|
||||
conv.convert(group_comment_);
|
||||
}
|
||||
}
|
||||
|
||||
// get info string
|
||||
std::string get_info(int width = 20)
|
||||
{
|
||||
// prepare printable trigger-time
|
||||
std::time_t tt = std::chrono::system_clock::to_time_t(trigger_time_);
|
||||
std::time_t att = std::chrono::system_clock::to_time_t(absolute_trigger_time_);
|
||||
|
||||
std::stringstream ss;
|
||||
ss<<std::setw(width)<<std::left<<"uuid:"<<uuid_<<"\n"
|
||||
<<std::setw(width)<<std::left<<"name:"<<name_<<"\n"
|
||||
<<std::setw(width)<<std::left<<"comment:"<<comment_<<"\n"
|
||||
<<std::setw(width)<<std::left<<"origin:"<<origin_<<"\n"
|
||||
<<std::setw(width)<<std::left<<"description:"<<text_<<"\n"
|
||||
<<std::setw(width)<<std::left<<"trigger-time-nt:"<<std::put_time(std::gmtime(&tt),"%FT%T")<<"\n"
|
||||
<<std::setw(width)<<std::left<<"trigger-time:"<<std::put_time(std::gmtime(&att),"%FT%T")<<"\n"
|
||||
<<std::setw(width)<<std::left<<"language-code:"<<language_code_<<"\n"
|
||||
<<std::setw(width)<<std::left<<"codepage:"<<codepage_<<"\n"
|
||||
<<std::setw(width)<<std::left<<"yname:"<<yname_<<"\n"
|
||||
<<std::setw(width)<<std::left<<"yunit:"<<yunit_<<"\n"
|
||||
<<std::setw(width)<<std::left<<"datatype:"<<datatp_<<"\n"
|
||||
<<std::setw(width)<<std::left<<"significant bits:"<<signbits_<<"\n"
|
||||
<<std::setw(width)<<std::left<<"buffer-offset:"<<buffer_offset_<<"\n"
|
||||
<<std::setw(width)<<std::left<<"buffer-size:"<<buffer_size_<<"\n"
|
||||
<<std::setw(width)<<std::left<<"add-time:"<<addtime_<<"\n"
|
||||
<<std::setw(width)<<std::left<<"xname:"<<xname_<<"\n"
|
||||
<<std::setw(width)<<std::left<<"xunit:"<<xunit_<<"\n"
|
||||
<<std::setw(width)<<std::left<<"xstepwidth:"<<xstepwidth_<<"\n"
|
||||
@@ -344,8 +539,8 @@ namespace imc
|
||||
<<std::setw(width)<<std::left<<"offset:"<<offset_<<"\n"
|
||||
<<std::setw(width)<<std::left<<"group:"<<"("<<group_index_<<","<<group_name_
|
||||
<<","<<group_comment_<<")"<<"\n"
|
||||
<<std::setw(width)<<std::left<<"ydata:"<<imc::joinvec<imc::datatype>(ydata_)<<"\n"
|
||||
<<std::setw(width)<<std::left<<"xdata:"<<imc::joinvec<double>(xdata_)<<"\n";
|
||||
<<std::setw(width)<<std::left<<"ydata:"<<imc::joinvec<imc::datatype>(ydata_,6,9,true)<<"\n"
|
||||
<<std::setw(width)<<std::left<<"xdata:"<<imc::joinvec<double>(xdata_,6,xprec_,true)<<"\n";
|
||||
// <<std::setw(width)<<std::left<<"aff. blocks:"<<chnenv_.get_json()<<"\n";
|
||||
return ss.str();
|
||||
}
|
||||
@@ -353,17 +548,26 @@ namespace imc
|
||||
// provide JSON string of metadata
|
||||
std::string get_json(bool include_data = false)
|
||||
{
|
||||
// prepare printable trigger-time
|
||||
std::time_t tt = std::chrono::system_clock::to_time_t(trigger_time_);
|
||||
std::time_t att = std::chrono::system_clock::to_time_t(absolute_trigger_time_);
|
||||
|
||||
std::stringstream ss;
|
||||
ss<<"{"<<"\"uuid\":\""<<uuid_
|
||||
<<"\",\"name\":\""<<name_
|
||||
<<"\",\"comment\":\""<<comment_
|
||||
<<"\",\"origin\":\""<<origin_
|
||||
<<"\",\"description\":\""<<text_
|
||||
<<"\",\"yname\":\""<<yname_
|
||||
<<"\",\"yunit\":\""<<yunit_
|
||||
<<"\",\"trigger-time-nt\":\""<<std::put_time(std::gmtime(&tt),"%FT%T")
|
||||
<<"\",\"trigger-time\":\""<<std::put_time(std::gmtime(&att),"%FT%T")
|
||||
<<"\",\"language-code\":\""<<language_code_
|
||||
<<"\",\"codepage\":\""<<codepage_
|
||||
<<"\",\"yname\":\""<<prepjsonstr(yname_)
|
||||
<<"\",\"yunit\":\""<<prepjsonstr(yunit_)
|
||||
<<"\",\"significantbits\":\""<<signbits_
|
||||
<<"\",\"xname\":\""<<xname_
|
||||
<<"\",\"xunit\":\""<<xunit_
|
||||
<<"\",\"addtime\":\""<<addtime_
|
||||
<<"\",\"xname\":\""<<prepjsonstr(xname_)
|
||||
<<"\",\"xunit\":\""<<prepjsonstr(xunit_)
|
||||
<<"\",\"xstepwidth\":\""<<xstepwidth_
|
||||
<<"\",\"xoffset\":\""<<xoffset_
|
||||
<<"\",\"group\":{"<<"\"index\":\""<<group_index_
|
||||
@@ -371,8 +575,8 @@ namespace imc
|
||||
<<"\",\"comment\":\""<<group_comment_<<"\""<<"}";
|
||||
if ( include_data )
|
||||
{
|
||||
ss<<",\"ydata\":"<<imc::joinvec<imc::datatype>(ydata_,0)
|
||||
<<",\"xdata\":"<<imc::joinvec<double>(xdata_,0);
|
||||
ss<<",\"ydata\":"<<imc::joinvec<imc::datatype>(ydata_,0,9,true)
|
||||
<<",\"xdata\":"<<imc::joinvec<double>(xdata_,0,xprec_,true);
|
||||
}
|
||||
// ss<<"\",\"aff. blocks\":\""<<chnenv_.get_json()
|
||||
ss<<"}";
|
||||
@@ -380,8 +584,27 @@ namespace imc
|
||||
return ss.str();
|
||||
}
|
||||
|
||||
// prepare string value for usage in JSON dump
|
||||
std::string prepjsonstr(std::string value)
|
||||
{
|
||||
std::stringstream ss;
|
||||
ss<<quoted(value);
|
||||
return strip_quotes(ss.str());
|
||||
}
|
||||
|
||||
// remove any leading or trailing double quotes
|
||||
std::string strip_quotes(std::string astring)
|
||||
{
|
||||
// head
|
||||
if ( astring.front() == '"' ) astring.erase(astring.begin()+0);
|
||||
// tail
|
||||
if ( astring.back() == '"' ) astring.erase(astring.end()-1);
|
||||
|
||||
return astring;
|
||||
}
|
||||
|
||||
// print channel
|
||||
void print(std::string filename, const char sep = ',', int width = 25)
|
||||
void print(std::string filename, const char sep = ',', int width = 25, int yprec = 9)
|
||||
{
|
||||
std::ofstream fou(filename);
|
||||
|
||||
@@ -402,13 +625,16 @@ namespace imc
|
||||
{
|
||||
if ( sep == ' ' )
|
||||
{
|
||||
fou<<std::setprecision(9)<<std::fixed
|
||||
fou<<std::setprecision(xprec_)<<std::fixed
|
||||
<<std::setw(width)<<std::left<<xdata_[i]
|
||||
<<std::setprecision(yprec)<<std::fixed
|
||||
<<std::setw(width)<<std::left<<ydata_[i]<<"\n";
|
||||
}
|
||||
else
|
||||
{
|
||||
fou<<xdata_[i]<<sep<<ydata_[i]<<"\n";
|
||||
fou<<std::setprecision(xprec_)<<std::fixed<<xdata_[i]
|
||||
<<sep
|
||||
<<std::setprecision(yprec)<<std::fixed<<ydata_[i]<<"\n";
|
||||
}
|
||||
}
|
||||
|
||||
|
@@ -1,6 +1,6 @@
|
||||
//---------------------------------------------------------------------------//
|
||||
|
||||
#ifndef IMCCONVRSION
|
||||
#ifndef IMCCONVERSION
|
||||
#define IMCCONVERSION
|
||||
|
||||
#include <vector>
|
||||
|
@@ -48,7 +48,7 @@ namespace imc
|
||||
public:
|
||||
datatype(): ubyte_(0), sbyte_(0),
|
||||
ushort_(0), sshort_(0),
|
||||
ulint_(0.0), slint_(0.0),
|
||||
ulint_(0), slint_(0),
|
||||
sfloat_(0.0), sdouble_(0.0),
|
||||
sdigital_(0),
|
||||
dtidx_(0) { };
|
||||
@@ -61,10 +61,10 @@ namespace imc
|
||||
datatype(imc_Slongint num): slint_(num), dtidx_(5) {};
|
||||
datatype(imc_float num): sfloat_(num), dtidx_(6) {};
|
||||
datatype(imc_double num): ubyte_(0), sbyte_(0), ushort_(0), sshort_(0),
|
||||
ulint_(0.0), slint_(0.0), sfloat_(0.0), sdouble_(num),
|
||||
ulint_(0), slint_(0), sfloat_(0.0), sdouble_(num),
|
||||
sdigital_(0), dtidx_(7) {};
|
||||
datatype(imc_digital num): ubyte_(0), sbyte_(0), ushort_(0), sshort_(0),
|
||||
ulint_(0.0), slint_(0.0), sfloat_(0.0), sdouble_(num),
|
||||
ulint_(0), slint_(0), sfloat_(0.0), sdouble_(num),
|
||||
sdigital_(num), dtidx_(10) {};
|
||||
|
||||
// identify type
|
||||
|
@@ -84,10 +84,12 @@ namespace imc
|
||||
// noncritical keys
|
||||
key(false,"NO","origin of data",1),
|
||||
key(false,"NT","timestamp of trigger",1),
|
||||
key(false,"NT","timestamp of trigger",2),
|
||||
key(false,"ND","(color) display properties",1),
|
||||
key(false,"NU","user defined key",1),
|
||||
key(false,"Np","property of channel",1),
|
||||
key(false,"NE","extraction rule for BUS channels",1)
|
||||
key(false,"NE","extraction rule for BUS channels",1),
|
||||
key(false,"NL","language info and code page",1)
|
||||
|
||||
};
|
||||
|
||||
|
@@ -500,22 +500,22 @@ namespace imc
|
||||
minute_ = std::stoi( get_parameter(buffer,¶meters[6]) );
|
||||
second_ = std::stod( get_parameter(buffer,¶meters[7]) );
|
||||
|
||||
time_t rawtime;
|
||||
struct tm ts;
|
||||
time(&rawtime);
|
||||
localtime_r(&rawtime,&ts);
|
||||
ts.tm_mday = day_;
|
||||
ts.tm_mon = month_-1;
|
||||
ts.tm_year = year_-1900;
|
||||
ts.tm_hour = hour_;
|
||||
ts.tm_min = minute_;
|
||||
ts.tm_sec = (int)second_;
|
||||
asctime_r(&ts,×tamp_[0]);
|
||||
// timestamp_ = std::to_string(year_) + std::string("-") + std::to_string(month_)
|
||||
// + std::string("-") + std::to_string(day_)
|
||||
// + std::string("T") + std::to_string(hour_)
|
||||
// + std::string(":") + std::to_string(minute_)
|
||||
// + std::string(":") + std::to_string(second_);
|
||||
//time_t rawtime;
|
||||
//struct tm ts;
|
||||
//time(&rawtime);
|
||||
//localtime_r(&rawtime,&ts);
|
||||
//ts.tm_mday = day_;
|
||||
//ts.tm_mon = month_-1;
|
||||
//ts.tm_year = year_-1900;
|
||||
//ts.tm_hour = hour_;
|
||||
//ts.tm_min = minute_;
|
||||
//ts.tm_sec = (int)second_;
|
||||
//asctime_r(&ts,×tamp_[0]);
|
||||
timestamp_ = std::to_string(year_) + std::string("-") + std::to_string(month_)
|
||||
+ std::string("-") + std::to_string(day_)
|
||||
+ std::string("T") + std::to_string(hour_)
|
||||
+ std::string(":") + std::to_string(minute_)
|
||||
+ std::string(":") + std::to_string(second_);
|
||||
}
|
||||
|
||||
// get info string
|
||||
|
@@ -58,6 +58,8 @@ namespace imc
|
||||
// open file and stream data into buffer
|
||||
void fill_buffer()
|
||||
{
|
||||
buffer_.clear();
|
||||
|
||||
// open file and put data in buffer
|
||||
try {
|
||||
std::ifstream fin(raw_file_.c_str(),std::ifstream::binary);
|
||||
@@ -76,6 +78,8 @@ namespace imc
|
||||
// parse all raw blocks in buffer
|
||||
void parse_blocks()
|
||||
{
|
||||
rawblocks_.clear();
|
||||
|
||||
// reset counter to identify computational complexity
|
||||
cplxcnt_ = 0;
|
||||
|
||||
@@ -123,20 +127,20 @@ namespace imc
|
||||
leng.push_back((char)*(it+pos));
|
||||
pos++;
|
||||
}
|
||||
unsigned long length = std::stoul(leng);
|
||||
unsigned long int length = std::stoul(leng);
|
||||
|
||||
// declare and initialize corresponding key and block
|
||||
// imc::key bkey( *(it+1)==imc::key_crit_ , newkey,
|
||||
// imc::keys.at(newkey).description_, version );
|
||||
imc::block blk(itkey,it-buffer_.begin(),
|
||||
it-buffer_.begin()+pos+1+length,
|
||||
imc::block blk(itkey,(unsigned long int)(it-buffer_.begin()),
|
||||
(unsigned long int)(it-buffer_.begin()+pos+1+length),
|
||||
raw_file_, &buffer_);
|
||||
|
||||
// add block to list
|
||||
rawblocks_.push_back(blk);
|
||||
|
||||
// skip the remaining block according to its length
|
||||
if ( it-buffer_.begin()+length < buffer_.size() )
|
||||
if ( (unsigned long int)(it-buffer_.begin()+length) < (unsigned long int)(buffer_.size()) )
|
||||
{
|
||||
std::advance(it,length);
|
||||
}
|
||||
@@ -190,6 +194,8 @@ namespace imc
|
||||
// generate map of blocks using their uuid
|
||||
void generate_block_map()
|
||||
{
|
||||
mapblocks_.clear();
|
||||
|
||||
for ( imc::block blk: rawblocks_ )
|
||||
{
|
||||
mapblocks_.insert( std::pair<std::string,imc::block>(blk.get_uuid(),blk) );
|
||||
@@ -199,6 +205,8 @@ namespace imc
|
||||
// generate channel "environments"
|
||||
void generate_channel_env()
|
||||
{
|
||||
channels_.clear();
|
||||
|
||||
// declare single channel environment
|
||||
imc::channel_env chnenv;
|
||||
chnenv.reset();
|
||||
@@ -207,7 +215,6 @@ namespace imc
|
||||
// BLOCK CORRESPONDENCE GOVERNED BY BLOCK ORDER IN BUFFER!!
|
||||
for ( imc::block blk: rawblocks_ )
|
||||
{
|
||||
|
||||
if ( blk.get_key().name_ == "CN" ) chnenv.CNuuid_ = blk.get_uuid();
|
||||
else if ( blk.get_key().name_ == "CD" ) chnenv.CDuuid_ = blk.get_uuid();
|
||||
else if ( blk.get_key().name_ == "CT" ) chnenv.CTuuid_ = blk.get_uuid();
|
||||
@@ -217,6 +224,7 @@ namespace imc
|
||||
else if ( blk.get_key().name_ == "CS" ) chnenv.CSuuid_ = blk.get_uuid();
|
||||
else if ( blk.get_key().name_ == "NT" ) chnenv.NTuuid_ = blk.get_uuid();
|
||||
else if ( blk.get_key().name_ == "NO" ) chnenv.NOuuid_ = blk.get_uuid();
|
||||
else if ( blk.get_key().name_ == "NL" ) chnenv.NLuuid_ = blk.get_uuid();
|
||||
|
||||
// check for currently associated channel
|
||||
if ( !chnenv.CNuuid_.empty() )
|
||||
@@ -228,13 +236,27 @@ namespace imc
|
||||
// provide UUID for channel
|
||||
chnenv.uuid_ = chnenv.CNuuid_;
|
||||
|
||||
// for multichannel data there may be multiple channels referring to
|
||||
// the same (final) CS block (in contrast to what the IMC software
|
||||
// documentation seems to suggest) resulting in all channels missing
|
||||
// a CS block except for the very last
|
||||
if ( chnenv.CSuuid_.empty() ) {
|
||||
for ( imc::block blkCS: rawblocks_ ) {
|
||||
if ( blkCS.get_key().name_ == "CS"
|
||||
&& blkCS.get_begin() > (unsigned long int)stol(chnenv.uuid_) ) {
|
||||
chnenv.CSuuid_ = blkCS.get_uuid();
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
// create channel object and add it to the map of channels
|
||||
channels_.insert( std::pair<std::string,imc::channel>
|
||||
(chnenv.CNuuid_,imc::channel(chnenv,&mapblocks_,&buffer_))
|
||||
);
|
||||
|
||||
// reset channel uuid
|
||||
chnenv.CNuuid_.clear();
|
||||
chnenv.reset();
|
||||
//chnenv.CNuuid_.clear();
|
||||
}
|
||||
}
|
||||
|
||||
@@ -246,13 +268,12 @@ namespace imc
|
||||
}
|
||||
}
|
||||
|
||||
|
||||
public:
|
||||
|
||||
// provide buffer size
|
||||
unsigned long int buffer_size()
|
||||
{
|
||||
return buffer_.size();
|
||||
return (unsigned long int)buffer_.size();
|
||||
}
|
||||
|
||||
// get blocks
|
||||
@@ -376,7 +397,7 @@ namespace imc
|
||||
std::filesystem::path pf = pd / filenam;
|
||||
|
||||
// and print the channel
|
||||
it->second.print(pf,sep);
|
||||
it->second.print(pf.u8string(),sep);
|
||||
}
|
||||
}
|
||||
|
||||
|
84
makefile
84
makefile
@@ -8,22 +8,25 @@ EXE = imctermite
|
||||
# directory names
|
||||
SRC = src/
|
||||
LIB = lib/
|
||||
CYT = cython/
|
||||
PYT = python/
|
||||
|
||||
# list headers
|
||||
# list headers and include directories
|
||||
HPP = $(wildcard $(LIB)/*.hpp)
|
||||
IPP = $(shell find $(LIB) -type f -name '*.hpp')
|
||||
KIB = $(shell find $(LIB) -type d)
|
||||
MIB = $(foreach dir,$(KIB),-I $(dir))
|
||||
|
||||
# choose compiler and its options
|
||||
CC = g++ -std=c++17
|
||||
#OPT = -O3 -Wall -mavx -mno-tbm -mf16c -mno-f16c
|
||||
OPT = -O3 -Wall -Werror -Wunused-variable -Wsign-compare
|
||||
OPT = -O3 -Wall -Wconversion -Wpedantic -Werror -Wunused-variable -Wsign-compare
|
||||
|
||||
# determine git version/commit and release tag
|
||||
GTAG := $(shell git tag | tail -n1)
|
||||
GTAG := $(shell git tag -l --sort=version:refname | tail -n1 | sed "s/$^v//g")
|
||||
GHSH := $(shell git rev-parse HEAD | head -c8)
|
||||
RTAG := v$(shell cat pip/setup.py | grep version | grep -oP "([0-9]\.){2}[0-9]")
|
||||
CTAG := v$(shell cat cython/setup.py | grep version | grep -oP "([0-9]\.){2}[0-9]")
|
||||
GVSN := $(shell cat python/VERSION | tr -d ' \n')
|
||||
|
||||
# current timestamp
|
||||
TMS = $(shell date +%Y%m%dT%H%M%S)
|
||||
|
||||
# define install location
|
||||
INST := /usr/local/bin
|
||||
@@ -31,50 +34,50 @@ INST := /usr/local/bin
|
||||
#-----------------------------------------------------------------------------#
|
||||
# C++ and CLI tool
|
||||
|
||||
# build exectuable
|
||||
$(EXE) : check-vtag $(RTAG) main.o
|
||||
# build executable
|
||||
$(EXE): check-tags $(GVSN) main.o
|
||||
$(CC) $(OPT) main.o -o $@
|
||||
|
||||
# build main.cpp and include git version/commit tag
|
||||
main.o : src/main.cpp $(HPP)
|
||||
main.o: src/main.cpp $(IPP)
|
||||
@cp $< $<.cpp
|
||||
@sed -i 's/TAGSTRING/$(GTAG)/g' $<.cpp
|
||||
@sed -i 's/HASHSTRING/$(GHSH)/g' $<.cpp
|
||||
$(CC) -c $(OPT) -I $(LIB) $<.cpp -o $@
|
||||
@sed -i 's/TIMESTAMPSTRING/$(TMS)/g' $<.cpp
|
||||
$(CC) -c $(OPT) $(MIB) $<.cpp -o $@
|
||||
@rm $<.cpp
|
||||
|
||||
install : $(EXE)
|
||||
install: $(EXE)
|
||||
cp $< $(INST)/
|
||||
|
||||
uninstall : $(INST)/$(EXE)
|
||||
uninstall: $(INST)/$(EXE)
|
||||
rm $<
|
||||
|
||||
cpp-clean :
|
||||
cpp-clean:
|
||||
rm -vf $(EXE)
|
||||
rm -vf *.o
|
||||
|
||||
#-----------------------------------------------------------------------------#
|
||||
# linter and code check
|
||||
# C++ linter
|
||||
|
||||
check-code:
|
||||
cppcheck --enable=all -I lib/ src/main.cpp
|
||||
|
||||
#-----------------------------------------------------------------------------#
|
||||
# check version consistency of git tags and version string in package.json
|
||||
# versions
|
||||
|
||||
$(GTAG) :
|
||||
$(GTAG):
|
||||
@echo "consistent versions check successful: building $(GTAG)"
|
||||
|
||||
check-vtag:
|
||||
@echo "git tag version: "$(GTAG)
|
||||
@echo "git commit hash: "$(GHSH)
|
||||
@echo "release version: "$(RTAG)
|
||||
@echo "module version: "$(CTAG)
|
||||
check-tags:
|
||||
@echo "latest git tag: $(GTAG)"
|
||||
@echo "latest git hash: $(GHSH)"
|
||||
@echo "python version: $(GVSN)"
|
||||
|
||||
#-----------------------------------------------------------------------------#
|
||||
# Docker
|
||||
|
||||
docker-build :
|
||||
docker-build:
|
||||
docker build ./ --tag imctermite:0.1
|
||||
|
||||
docker-run:
|
||||
@@ -83,28 +86,29 @@ docker-run:
|
||||
#-----------------------------------------------------------------------------#
|
||||
# python
|
||||
|
||||
cython-build : check-vtag $(CTAG) $(CYT)setup.py $(CYT)imc_termite.pxd $(CYT)py_imc_termite.pyx $(HPP)
|
||||
python3 $(CYT)setup.py build_ext --inplace
|
||||
cp -v imc_termite.cpython-*.so $(PYT)
|
||||
python-build: check-tags $(GVSN)
|
||||
make -C python/ build-inplace
|
||||
cp python/IMCtermite*.so ./ -v
|
||||
|
||||
cython-install : check-vtag $(CTAG) $(CYT)setup.py $(CYT)imc_termite.pxd $(CYT)py_imc_termite.pyx $(HPP)
|
||||
python3 $(CYT)setup.py install --record files_imctermite.txt
|
||||
python-clean:
|
||||
make -C python/ clean
|
||||
rm -vf IMCtermite*.so
|
||||
|
||||
cython-clean :
|
||||
rm -vf imc_termite.cpython-*.so
|
||||
rm -vf $(PYT)imc_termite.cpython-*.so
|
||||
rm -rvf build/
|
||||
|
||||
#-----------------------------------------------------------------------------#
|
||||
# pip
|
||||
|
||||
pip-release: check-vtag $(RTAG) cython-build
|
||||
cd ./pip/ && make publish
|
||||
python-test:
|
||||
PYTHONPATH=./ python python/examples/usage.py
|
||||
|
||||
#-----------------------------------------------------------------------------#
|
||||
# clean
|
||||
|
||||
clean: cpp-clean cython-clean
|
||||
cd ./pip/ && make clean
|
||||
clean: cpp-clean python-clean
|
||||
|
||||
#-----------------------------------------------------------------------------#
|
||||
# github actions
|
||||
|
||||
github-action-lint: .github/workflows/pypi-deploy.yml
|
||||
actionlint $<
|
||||
|
||||
# for reference, see:
|
||||
# https://github.com/rhysd/actionlint
|
||||
|
||||
#-----------------------------------------------------------------------------#
|
||||
|
@@ -1,133 +0,0 @@
|
||||
#-----------------------------------------------------------------------------#
|
||||
|
||||
import argparse
|
||||
import os
|
||||
|
||||
#-----------------------------------------------------------------------------#
|
||||
|
||||
parser = argparse.ArgumentParser(description='List all source dependencies')
|
||||
#parser.add_argument('pathToRepo',type=str,help='path of source repository')
|
||||
parser.add_argument('mainSource',type=str,help='main source file')
|
||||
parser.add_argument('depFile',type=str,help='file listing all dependencies')
|
||||
args = parser.parse_args()
|
||||
|
||||
libpaths = ["/home/mario/Desktop/arrow/cpp/src/",
|
||||
"/home/mario/Desktop/arrow/cpp/thrift_ep-install/include/",
|
||||
"/home/mario/Desktop/arrow/cpp/boost_ep-prefix/src/boost_ep/"]
|
||||
|
||||
#-----------------------------------------------------------------------------#
|
||||
|
||||
def find_dependencies(srcfile, recdepth, cdeplist) :
|
||||
"""
|
||||
Given a source file and its dependencies in the given repository path
|
||||
list all further dependencies recursively
|
||||
|
||||
Args:
|
||||
srcfile (string): path/name of source file
|
||||
recdepth (integer): current recursion depth
|
||||
cdeplist (list): current list of dependencies
|
||||
Return:
|
||||
deps (list): list of source files in repository, the source file depends on
|
||||
"""
|
||||
# define indentation to visual recursion
|
||||
indent = recdepth*(" ")
|
||||
|
||||
print("\n" + indent + "find_dependencies:"
|
||||
+ "\n" + indent + "1: " + srcfile
|
||||
+ "\n" + indent + "2: " + str(recdepth)
|
||||
+ "\n" + indent + "3: " + str(len(cdeplist)) + "\n")
|
||||
|
||||
# show dependencies so far
|
||||
#print(cdeplist)
|
||||
|
||||
# generate dependencies by means of g++
|
||||
libdeps = (" -I ").join(libpaths)
|
||||
cmd = "g++ -c -MMD " + srcfile + " -I " + libdeps
|
||||
print(indent + cmd )
|
||||
os.system(cmd)
|
||||
|
||||
# open dependency file and extract list of sources
|
||||
basename = srcfile.split('/')[-1].split('.')[0]
|
||||
depfile = basename + '.d'
|
||||
print(indent + "reading dependency file " + depfile)
|
||||
with open(depfile,'r') as fin :
|
||||
depslist = fin.readlines()
|
||||
|
||||
# delete dependencies and object files
|
||||
os.system("rm " + basename + ".d")
|
||||
os.system("rm " + basename + ".o")
|
||||
|
||||
# remove first line
|
||||
depslist = depslist[1:]
|
||||
|
||||
# delete leading space and trailing backslash
|
||||
depslistcl = [dep.lstrip(' ').rstrip(' \\\n') for dep in depslist]
|
||||
|
||||
# collect dependencies
|
||||
newdeps = []
|
||||
|
||||
# check all dependencies recursively and collect further dependencies
|
||||
count = 0
|
||||
for dep in depslistcl :
|
||||
# append source itself to list
|
||||
if dep not in cdeplist :
|
||||
print(indent + "adding dependency " + dep)
|
||||
newdeps.append(dep)
|
||||
count = count + 1
|
||||
print(indent + "=> added " + str(count) + "/" + str(len(depslistcl)) )
|
||||
|
||||
# check recursion depth
|
||||
if recdepth < 20 :
|
||||
# check all dependencies of every single dependency
|
||||
for dep in depslistcl :
|
||||
# try to find corresponding *.cc, (*.cpp) file
|
||||
depcc = dep.split('.')[0] + '.cc'
|
||||
print(indent + "checking for " + depcc)
|
||||
if os.path.exists(depcc) :
|
||||
if depcc not in cdeplist and depcc not in newdeps :
|
||||
# add file itself as dependency
|
||||
newdeps.append(depcc)
|
||||
# find dependencies of single source
|
||||
newrecdeps = find_dependencies(depcc,recdepth+1,cdeplist+newdeps)
|
||||
# append to list
|
||||
for el in newrecdeps :
|
||||
if el not in newdeps :
|
||||
newdeps.append(el)
|
||||
else :
|
||||
print(indent + "already in list")
|
||||
else :
|
||||
print(indent + "does not exist")
|
||||
|
||||
print("\n")
|
||||
|
||||
# provide list of dependencies
|
||||
return newdeps
|
||||
|
||||
#-----------------------------------------------------------------------------#
|
||||
|
||||
if __name__== "__main__":
|
||||
|
||||
print("\nCLI arguments:\n" + str(args) + "\n")
|
||||
|
||||
# collect list of dependencies
|
||||
deps = []
|
||||
|
||||
# start recursion with given source file
|
||||
deps = find_dependencies(args.mainSource,0,[])
|
||||
|
||||
print("\nfinal list of dependencies: (" + str(len(deps)) + ")\n")
|
||||
print(deps)
|
||||
print("\n")
|
||||
|
||||
# remove any duplicates
|
||||
depsuni = set(deps)
|
||||
print("\nfinal set of dependencies: (" + str(len(depsuni)) + ")\n")
|
||||
print(depsuni)
|
||||
print("\n")
|
||||
|
||||
# write list of dependencies
|
||||
with open(args.depFile,'w') as fout :
|
||||
for el in depsuni :
|
||||
fout.write(str(el) + '\n')
|
||||
|
||||
#-----------------------------------------------------------------------------#
|
@@ -1,22 +0,0 @@
|
||||
#-----------------------------------------------------------------------------#
|
||||
|
||||
from pathlib import Path
|
||||
|
||||
# find source files
|
||||
srcpaths = Path("src/").rglob('*.cc')
|
||||
deps =[ str(path) for path in srcpaths ]
|
||||
|
||||
print(deps)
|
||||
|
||||
with open('makefileobj','w') as fout :
|
||||
for el in deps :
|
||||
basnam = el.split('/')[-1]
|
||||
print(str(el) + " : " + str(basnam) + " : " + str(basnam.split('.')[1]))
|
||||
if basnam.split('.')[1] == 'cc' :
|
||||
objfile = 'bin/' + basnam.replace('.cc','.o')
|
||||
fout.write(objfile + " : " + el + "\n")
|
||||
fout.write("\t" + "$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@\n")
|
||||
fout.write("\n")
|
||||
|
||||
|
||||
#-----------------------------------------------------------------------------#
|
@@ -1,356 +0,0 @@
|
||||
#-----------------------------------------------------------------------------#
|
||||
|
||||
CPP := g++ -std=c++14
|
||||
CPPFLAGS := -Woverflow -Wpedantic -Wextra -Waddress -Waligned-new -Walloc-zero
|
||||
|
||||
SRC := src/
|
||||
BIN := bin/
|
||||
|
||||
LIBS := -I src/src/ -I src/thrift_ep-install/include/ -I src/boost_ep-prefix/src/boost_ep/
|
||||
|
||||
#-----------------------------------------------------------------------------#
|
||||
# prepare source
|
||||
#
|
||||
# before: $ cd arrow/cpp/ and compile relevant sources by
|
||||
# $ cmake . -D ARROW_PARQUET=ON -D PARQUET_BUILD_EXAMPLES=ON -D ARROW_WITH_SNAPPY=ON
|
||||
# $ cmake .. -D ARROW_PARQUET=ON ARROW_BUILD_EXAMPLES=ON
|
||||
|
||||
lib :
|
||||
cmake . -D ARROW_WITH_BROTLI=ON -D ARROW_WITH_BZ2=ON -D ARROW_WITH_LZ4=ON -D ARROW_WITH_SNAPPY=ON -D ARROW_WITH_ZLIB=ON -D ARROW_PARQUET=ON -D ARROW_PYTHON=ON
|
||||
|
||||
# cp-src : deps.log
|
||||
# ./src_copy.sh
|
||||
|
||||
deps.log :
|
||||
python3 generate_deps.py reader-writer.cc $@
|
||||
|
||||
SRC := $(shell find $(SRC) -name '*.cc')
|
||||
|
||||
# OBJ := $(apprefix obj/, $(SRC:%.cc=%.o))
|
||||
OBJ := $(addprefix $(BIN),$(notdir $(SRC:%.cc=%.o)))
|
||||
|
||||
check :
|
||||
@echo $(SRC)
|
||||
@echo $(OBJ)
|
||||
|
||||
# vpath %.cc src/
|
||||
|
||||
reader-writer-example : reader-writer.cc $(OBJ) bin/utilmemory.o
|
||||
$(CPP) $(CPPFLAGS) $< $(LIBS) -o $@ $(OBJ) bin/utilmemory.o
|
||||
|
||||
# $(OBJ) : $(SRC)
|
||||
# $(CPP) $(OPT) -c $< -o $@ -I src/src/
|
||||
#
|
||||
# $(BIN)%.o : $(SRC)
|
||||
# $(CPP) $(OPT) -c $< -I src/src/ -o $@
|
||||
|
||||
clean-obj :
|
||||
rm -f $(OBJ)
|
||||
|
||||
# => do build with cmake like here
|
||||
# https://arrow.apache.org/docs/developers/python.html#build-and-test
|
||||
|
||||
#-----------------------------------------------------------------------------#
|
||||
|
||||
bin/type.o : src/src/arrow/type.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/result.o : src/src/arrow/result.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/builder.o : src/src/arrow/builder.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/tensor.o : src/src/arrow/tensor.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/table.o : src/src/arrow/table.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/extension_type.o : src/src/arrow/extension_type.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/device.o : src/src/arrow/device.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/memory_pool.o : src/src/arrow/memory_pool.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/datum.o : src/src/arrow/datum.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/record_batch.o : src/src/arrow/record_batch.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/compare.o : src/src/arrow/compare.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/visitor.o : src/src/arrow/visitor.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/chunked_array.o : src/src/arrow/chunked_array.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/status.o : src/src/arrow/status.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/pretty_print.o : src/src/arrow/pretty_print.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/sparse_tensor.o : src/src/arrow/sparse_tensor.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/buffer.o : src/src/arrow/buffer.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/scalar.o : src/src/arrow/scalar.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/string.o : src/src/arrow/util/string.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/utilmemory.o : src/src/arrow/util/memory.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/future.o : src/src/arrow/util/future.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/iterator.o : src/src/arrow/util/iterator.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/compression.o : src/src/arrow/util/compression.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/utf8.o : src/src/arrow/util/utf8.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/time.o : src/src/arrow/util/time.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/cpu_info.o : src/src/arrow/util/cpu_info.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/thread_pool.o : src/src/arrow/util/thread_pool.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/bit_util.o : src/src/arrow/util/bit_util.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/logging.o : src/src/arrow/util/logging.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/basic_decimal.o : src/src/arrow/util/basic_decimal.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/decimal.o : src/src/arrow/util/decimal.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/bit_block_counter.o : src/src/arrow/util/bit_block_counter.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/key_value_metadata.o : src/src/arrow/util/key_value_metadata.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/int_util.o : src/src/arrow/util/int_util.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/io_util.o : src/src/arrow/util/io_util.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/bitmap_ops.o : src/src/arrow/util/bitmap_ops.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/bitmap_builders.o : src/src/arrow/util/bitmap_builders.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/bit_run_reader.o : src/src/arrow/util/bit_run_reader.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/value_parsing.o : src/src/arrow/util/value_parsing.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/string_builder.o : src/src/arrow/util/string_builder.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/formatting.o : src/src/arrow/util/formatting.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/array_primitive.o : src/src/arrow/array/array_primitive.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/array_dict.o : src/src/arrow/array/array_dict.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/builder_binary.o : src/src/arrow/array/builder_binary.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/builder_union.o : src/src/arrow/array/builder_union.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/concatenate.o : src/src/arrow/array/concatenate.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/array_nested.o : src/src/arrow/array/array_nested.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/array_decimal.o : src/src/arrow/array/array_decimal.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/builder_primitive.o : src/src/arrow/array/builder_primitive.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/data.o : src/src/arrow/array/data.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/diff.o : src/src/arrow/array/diff.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/builder_nested.o : src/src/arrow/array/builder_nested.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/builder_decimal.o : src/src/arrow/array/builder_decimal.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/builder_dict.o : src/src/arrow/array/builder_dict.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/array_binary.o : src/src/arrow/array/array_binary.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/builder_adaptive.o : src/src/arrow/array/builder_adaptive.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/array_base.o : src/src/arrow/array/array_base.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/validate.o : src/src/arrow/array/validate.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/builder_base.o : src/src/arrow/array/builder_base.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/util.o : src/src/arrow/array/util.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/caching.o : src/src/arrow/io/caching.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/memory.o : src/src/arrow/io/memory.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/interfaces.o : src/src/arrow/io/interfaces.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/buffered.o : src/src/arrow/io/buffered.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/file.o : src/src/arrow/io/file.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/strtod.o : src/src/arrow/vendored/double-conversion/strtod.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/bignum.o : src/src/arrow/vendored/double-conversion/bignum.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/fixed-dtoa.o : src/src/arrow/vendored/double-conversion/fixed-dtoa.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/fast-dtoa.o : src/src/arrow/vendored/double-conversion/fast-dtoa.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/diy-fp.o : src/src/arrow/vendored/double-conversion/diy-fp.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/double-conversion.o : src/src/arrow/vendored/double-conversion/double-conversion.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/bignum-dtoa.o : src/src/arrow/vendored/double-conversion/bignum-dtoa.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/cached-powers.o : src/src/arrow/vendored/double-conversion/cached-powers.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/api_aggregate.o : src/src/arrow/compute/api_aggregate.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/exec.o : src/src/arrow/compute/exec.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/kernel.o : src/src/arrow/compute/kernel.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/registry.o : src/src/arrow/compute/registry.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/function.o : src/src/arrow/compute/function.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/cast.o : src/src/arrow/compute/cast.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/api_vector.o : src/src/arrow/compute/api_vector.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/api_scalar.o : src/src/arrow/compute/api_scalar.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/codegen_internal.o : src/src/arrow/compute/kernels/codegen_internal.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/column_scanner.o : src/src/parquet/column_scanner.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/statistics.o : src/src/parquet/statistics.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/internal_file_decryptor.o : src/src/parquet/internal_file_decryptor.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/column_writer.o : src/src/parquet/column_writer.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/encryption.o : src/src/parquet/encryption.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/file_reader.o : src/src/parquet/file_reader.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/properties.o : src/src/parquet/properties.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/encryption_internal.o : src/src/parquet/encryption_internal.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/internal_file_encryptor.o : src/src/parquet/internal_file_encryptor.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/types.o : src/src/parquet/types.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/encoding.o : src/src/parquet/encoding.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/metadata.o : src/src/parquet/metadata.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/printer.o : src/src/parquet/printer.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/level_conversion.o : src/src/parquet/level_conversion.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/deprecated_io.o : src/src/parquet/deprecated_io.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/file_writer.o : src/src/parquet/file_writer.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/schema.o : src/src/parquet/schema.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/platform.o : src/src/parquet/platform.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/column_reader.o : src/src/parquet/column_reader.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
@@ -1,96 +0,0 @@
|
||||
#-----------------------------------------------------------------------------#
|
||||
|
||||
PARQUETDIR := /home/mario/Desktop/Record_Evolution/parquet-cpp
|
||||
ARROWDIR := /home/mario/Desktop/Record_Evolution/arrow/cpp/src
|
||||
|
||||
CPP := g++ -std=c++14
|
||||
OPT := -Wall -Woverflow -Wpedantic -Wextra -Waddress -Waligned-new -Walloc-zero
|
||||
|
||||
prepare : collect_parquet modify_parquet collect_arrow modify_arrow
|
||||
|
||||
collect_parquet :
|
||||
cp -r $(PARQUETDIR)/src/parquet ./
|
||||
cp $(PARQUETDIR)/examples/low-level-api/reader_writer.h ./
|
||||
cp $(PARQUETDIR)/examples/low-level-api/reader-writer.cc ./
|
||||
|
||||
modify_parquet :
|
||||
cp parquet/parquet_version.h.in parquet/parquet_version.h
|
||||
sed -i 's/ReadableFileInterface/ReadWriteFileInterface/g' parquet/util/memory.h
|
||||
sed -i 's/ReadableFileInterface/ReadWriteFileInterface/g' parquet/file_reader.h
|
||||
sed -i 's/arrow::Codec/arrow::util::Codec/g' parquet/util/memory.h
|
||||
sed -i 's/valid_bits_writer/valid_bits_offset/g' parquet/column_reader.h
|
||||
|
||||
collect_arrow :
|
||||
cp -r $(ARROWDIR)/arrow ./
|
||||
|
||||
modify_arrow :
|
||||
cp arrow/util/bit_util.h arrow/util/bit-util.h
|
||||
|
||||
collect_test :
|
||||
cp $(PARQUETDIR)/examples/low-level-api/reader-writer.cc ./
|
||||
|
||||
subst :
|
||||
sed -i 's/#include \"arrow\//\/\/#include \"arrow/g' parquet/properties.h
|
||||
|
||||
test :
|
||||
$(CPP) $(OPT) -I$(PWD) reader-writer.cc
|
||||
|
||||
clean :
|
||||
rm -r parquet/ arrow/
|
||||
rm reader-writer.cc reader_writer.h
|
||||
|
||||
#-----------------------------------------------------------------------------#
|
||||
|
||||
# choose shell
|
||||
SHELL:=/bin/bash
|
||||
|
||||
SRC = reader-writer
|
||||
|
||||
# specify path of cloned directory
|
||||
ARROWGIT := /home/mario/Desktop/Record_Evolution/arrow
|
||||
|
||||
filewriter : parquet/file_writer.cc
|
||||
$(CPP) -c $(OPT) $<
|
||||
|
||||
# build executable (and generate dependency file)
|
||||
readwrite : reader-writer.cc
|
||||
$(CPP) $(OPT) -MMD $< -I ./
|
||||
|
||||
# generate dependency file
|
||||
$(SRC).d : $(SRC).cc
|
||||
$(CPP) -c -MMD $< -I ./ -I $(ARROWGIT)/cpp/src/
|
||||
|
||||
# extract source dependencies
|
||||
extract-dep : $(SRC).d
|
||||
@# extract relevant dependencies
|
||||
cat $< | sed 's/ /\n/g' | awk 'NF' | grep -v '\\' | grep '\/' > deps.log
|
||||
cat deps.log | sed ':a;N;$!ba;s/\n/ /g' > headers.log
|
||||
cat headers.log | sed 's/.h$$/.cc/g' > sources.log
|
||||
@# copy required sources
|
||||
mkdir -p temp/
|
||||
cp --parents `cat headers.log` temp/
|
||||
cp --parents `cat sources.log` temp/ 2>/dev/null
|
||||
mv temp$(ARROWGIT)/cpp/src/* ./
|
||||
rm -r temp
|
||||
|
||||
clean-dep :
|
||||
rm -f deps.log headers.log sources.log $(SRC).d
|
||||
|
||||
#-----------------------------------------------------------------------------#
|
||||
# only use more recent and up to date repository arrow.git
|
||||
|
||||
# build arrow shared/static libraries
|
||||
build :
|
||||
cd arrow/cpp
|
||||
# cmake -LA to show all options
|
||||
cmake . -D ARROW_PARQUET=ON #ARROW_ARMV8_ARCH=armv8-a
|
||||
make
|
||||
|
||||
example :
|
||||
cd arrow/cpp/examples/parquet/low-level-api/
|
||||
g++ reader-writer.cc -I. -I../../../src/ -L../../../../cpp/build/release/ -larrow -lparquet
|
||||
|
||||
# set environment variable LD_LIBRARY_PATH=../../../../cpp/build/release/ before launching executable
|
||||
|
||||
|
||||
#------------------------------------------------------------------------------------#
|
@@ -1,303 +0,0 @@
|
||||
bin/type.o : src/src/arrow/type.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/result.o : src/src/arrow/result.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/builder.o : src/src/arrow/builder.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/tensor.o : src/src/arrow/tensor.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/table.o : src/src/arrow/table.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/extension_type.o : src/src/arrow/extension_type.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/device.o : src/src/arrow/device.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/memory_pool.o : src/src/arrow/memory_pool.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/datum.o : src/src/arrow/datum.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/record_batch.o : src/src/arrow/record_batch.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/compare.o : src/src/arrow/compare.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/visitor.o : src/src/arrow/visitor.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/chunked_array.o : src/src/arrow/chunked_array.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/status.o : src/src/arrow/status.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/pretty_print.o : src/src/arrow/pretty_print.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/sparse_tensor.o : src/src/arrow/sparse_tensor.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/buffer.o : src/src/arrow/buffer.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/scalar.o : src/src/arrow/scalar.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/string.o : src/src/arrow/util/string.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/memory.o : src/src/arrow/util/memory.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/future.o : src/src/arrow/util/future.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/iterator.o : src/src/arrow/util/iterator.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/compression.o : src/src/arrow/util/compression.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/utf8.o : src/src/arrow/util/utf8.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/time.o : src/src/arrow/util/time.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/cpu_info.o : src/src/arrow/util/cpu_info.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/thread_pool.o : src/src/arrow/util/thread_pool.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/bit_util.o : src/src/arrow/util/bit_util.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/logging.o : src/src/arrow/util/logging.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/basic_decimal.o : src/src/arrow/util/basic_decimal.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/decimal.o : src/src/arrow/util/decimal.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/bit_block_counter.o : src/src/arrow/util/bit_block_counter.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/key_value_metadata.o : src/src/arrow/util/key_value_metadata.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/int_util.o : src/src/arrow/util/int_util.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/io_util.o : src/src/arrow/util/io_util.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/bitmap_ops.o : src/src/arrow/util/bitmap_ops.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/bitmap_builders.o : src/src/arrow/util/bitmap_builders.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/bit_run_reader.o : src/src/arrow/util/bit_run_reader.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/value_parsing.o : src/src/arrow/util/value_parsing.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/string_builder.o : src/src/arrow/util/string_builder.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/formatting.o : src/src/arrow/util/formatting.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/array_primitive.o : src/src/arrow/array/array_primitive.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/array_dict.o : src/src/arrow/array/array_dict.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/builder_binary.o : src/src/arrow/array/builder_binary.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/builder_union.o : src/src/arrow/array/builder_union.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/concatenate.o : src/src/arrow/array/concatenate.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/array_nested.o : src/src/arrow/array/array_nested.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/array_decimal.o : src/src/arrow/array/array_decimal.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/builder_primitive.o : src/src/arrow/array/builder_primitive.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/data.o : src/src/arrow/array/data.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/diff.o : src/src/arrow/array/diff.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/builder_nested.o : src/src/arrow/array/builder_nested.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/builder_decimal.o : src/src/arrow/array/builder_decimal.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/builder_dict.o : src/src/arrow/array/builder_dict.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/array_binary.o : src/src/arrow/array/array_binary.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/builder_adaptive.o : src/src/arrow/array/builder_adaptive.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/array_base.o : src/src/arrow/array/array_base.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/validate.o : src/src/arrow/array/validate.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/builder_base.o : src/src/arrow/array/builder_base.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/util.o : src/src/arrow/array/util.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/caching.o : src/src/arrow/io/caching.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/memory.o : src/src/arrow/io/memory.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/interfaces.o : src/src/arrow/io/interfaces.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/buffered.o : src/src/arrow/io/buffered.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/file.o : src/src/arrow/io/file.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/strtod.o : src/src/arrow/vendored/double-conversion/strtod.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/bignum.o : src/src/arrow/vendored/double-conversion/bignum.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/fixed-dtoa.o : src/src/arrow/vendored/double-conversion/fixed-dtoa.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/fast-dtoa.o : src/src/arrow/vendored/double-conversion/fast-dtoa.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/diy-fp.o : src/src/arrow/vendored/double-conversion/diy-fp.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/double-conversion.o : src/src/arrow/vendored/double-conversion/double-conversion.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/bignum-dtoa.o : src/src/arrow/vendored/double-conversion/bignum-dtoa.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/cached-powers.o : src/src/arrow/vendored/double-conversion/cached-powers.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/api_aggregate.o : src/src/arrow/compute/api_aggregate.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/exec.o : src/src/arrow/compute/exec.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/kernel.o : src/src/arrow/compute/kernel.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/registry.o : src/src/arrow/compute/registry.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/function.o : src/src/arrow/compute/function.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/cast.o : src/src/arrow/compute/cast.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/api_vector.o : src/src/arrow/compute/api_vector.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/api_scalar.o : src/src/arrow/compute/api_scalar.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/codegen_internal.o : src/src/arrow/compute/kernels/codegen_internal.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/column_scanner.o : src/src/parquet/column_scanner.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/statistics.o : src/src/parquet/statistics.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/internal_file_decryptor.o : src/src/parquet/internal_file_decryptor.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/column_writer.o : src/src/parquet/column_writer.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/encryption.o : src/src/parquet/encryption.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/file_reader.o : src/src/parquet/file_reader.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/properties.o : src/src/parquet/properties.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/encryption_internal.o : src/src/parquet/encryption_internal.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/internal_file_encryptor.o : src/src/parquet/internal_file_encryptor.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/types.o : src/src/parquet/types.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/encoding.o : src/src/parquet/encoding.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/metadata.o : src/src/parquet/metadata.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/printer.o : src/src/parquet/printer.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/level_conversion.o : src/src/parquet/level_conversion.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/deprecated_io.o : src/src/parquet/deprecated_io.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/file_writer.o : src/src/parquet/file_writer.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/schema.o : src/src/parquet/schema.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/platform.o : src/src/parquet/platform.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
||||
bin/column_reader.o : src/src/parquet/column_reader.cc
|
||||
$(CPP) $(CPPFLAGS) -c $< $(LIBS) -o $@
|
||||
|
@@ -1,413 +0,0 @@
|
||||
// Licensed to the Apache Software Foundation (ASF) under one
|
||||
// or more contributor license agreements. See the NOTICE file
|
||||
// distributed with this work for additional information
|
||||
// regarding copyright ownership. The ASF licenses this file
|
||||
// to you under the Apache License, Version 2.0 (the
|
||||
// "License"); you may not use this file except in compliance
|
||||
// with the License. You may obtain a copy of the License at
|
||||
//
|
||||
// http://www.apache.org/licenses/LICENSE-2.0
|
||||
//
|
||||
// Unless required by applicable law or agreed to in writing,
|
||||
// software distributed under the License is distributed on an
|
||||
// "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
|
||||
// KIND, either express or implied. See the License for the
|
||||
// specific language governing permissions and limitations
|
||||
// under the License.
|
||||
|
||||
#include <cassert>
|
||||
#include <fstream>
|
||||
#include <iostream>
|
||||
#include <memory>
|
||||
|
||||
#include "reader_writer.h"
|
||||
|
||||
/*
|
||||
* This example describes writing and reading Parquet Files in C++ and serves as a
|
||||
* reference to the API.
|
||||
* The file contains all the physical data types supported by Parquet.
|
||||
* This example uses the RowGroupWriter API that supports writing RowGroups optimized for
|
||||
*memory consumption
|
||||
**/
|
||||
|
||||
/* Parquet is a structured columnar file format
|
||||
* Parquet File = "Parquet data" + "Parquet Metadata"
|
||||
* "Parquet data" is simply a vector of RowGroups. Each RowGroup is a batch of rows in a
|
||||
* columnar layout
|
||||
* "Parquet Metadata" contains the "file schema" and attributes of the RowGroups and their
|
||||
* Columns
|
||||
* "file schema" is a tree where each node is either a primitive type (leaf nodes) or a
|
||||
* complex (nested) type (internal nodes)
|
||||
* For specific details, please refer the format here:
|
||||
* https://github.com/apache/parquet-format/blob/master/LogicalTypes.md
|
||||
**/
|
||||
|
||||
constexpr int NUM_ROWS_PER_ROW_GROUP = 500;
|
||||
const char PARQUET_FILENAME[] = "parquet_cpp_example.parquet";
|
||||
|
||||
int main(int argc, char** argv) {
|
||||
/**********************************************************************************
|
||||
PARQUET WRITER EXAMPLE
|
||||
**********************************************************************************/
|
||||
// parquet::REQUIRED fields do not need definition and repetition level values
|
||||
// parquet::OPTIONAL fields require only definition level values
|
||||
// parquet::REPEATED fields require both definition and repetition level values
|
||||
try {
|
||||
// Create a local file output stream instance.
|
||||
using FileClass = ::arrow::io::FileOutputStream;
|
||||
std::shared_ptr<FileClass> out_file;
|
||||
PARQUET_ASSIGN_OR_THROW(out_file, FileClass::Open(PARQUET_FILENAME));
|
||||
|
||||
// Setup the parquet schema
|
||||
std::shared_ptr<GroupNode> schema = SetupSchema();
|
||||
|
||||
// Add writer properties
|
||||
parquet::WriterProperties::Builder builder;
|
||||
builder.compression(parquet::Compression::UNCOMPRESSED);
|
||||
std::shared_ptr<parquet::WriterProperties> props = builder.build();
|
||||
|
||||
// Create a ParquetFileWriter instance
|
||||
std::shared_ptr<parquet::ParquetFileWriter> file_writer =
|
||||
parquet::ParquetFileWriter::Open(out_file, schema, props);
|
||||
|
||||
// Append a RowGroup with a specific number of rows.
|
||||
parquet::RowGroupWriter* rg_writer = file_writer->AppendRowGroup();
|
||||
|
||||
// Write the Bool column
|
||||
parquet::BoolWriter* bool_writer =
|
||||
static_cast<parquet::BoolWriter*>(rg_writer->NextColumn());
|
||||
for (int i = 0; i < NUM_ROWS_PER_ROW_GROUP; i++) {
|
||||
bool value = ((i % 2) == 0) ? true : false;
|
||||
bool_writer->WriteBatch(1, nullptr, nullptr, &value);
|
||||
}
|
||||
|
||||
// Write the Int32 column
|
||||
parquet::Int32Writer* int32_writer =
|
||||
static_cast<parquet::Int32Writer*>(rg_writer->NextColumn());
|
||||
for (int i = 0; i < NUM_ROWS_PER_ROW_GROUP; i++) {
|
||||
int32_t value = i;
|
||||
int32_writer->WriteBatch(1, nullptr, nullptr, &value);
|
||||
}
|
||||
|
||||
// Write the Int64 column. Each row has repeats twice.
|
||||
parquet::Int64Writer* int64_writer =
|
||||
static_cast<parquet::Int64Writer*>(rg_writer->NextColumn());
|
||||
for (int i = 0; i < 2 * NUM_ROWS_PER_ROW_GROUP; i++) {
|
||||
int64_t value = i * 1000 * 1000;
|
||||
value *= 1000 * 1000;
|
||||
int16_t definition_level = 1;
|
||||
int16_t repetition_level = 0;
|
||||
if ((i % 2) == 0) {
|
||||
repetition_level = 1; // start of a new record
|
||||
}
|
||||
int64_writer->WriteBatch(1, &definition_level, &repetition_level, &value);
|
||||
}
|
||||
|
||||
// Write the INT96 column.
|
||||
parquet::Int96Writer* int96_writer =
|
||||
static_cast<parquet::Int96Writer*>(rg_writer->NextColumn());
|
||||
for (int i = 0; i < NUM_ROWS_PER_ROW_GROUP; i++) {
|
||||
parquet::Int96 value;
|
||||
value.value[0] = i;
|
||||
value.value[1] = i + 1;
|
||||
value.value[2] = i + 2;
|
||||
int96_writer->WriteBatch(1, nullptr, nullptr, &value);
|
||||
}
|
||||
|
||||
// Write the Float column
|
||||
parquet::FloatWriter* float_writer =
|
||||
static_cast<parquet::FloatWriter*>(rg_writer->NextColumn());
|
||||
for (int i = 0; i < NUM_ROWS_PER_ROW_GROUP; i++) {
|
||||
float value = static_cast<float>(i) * 1.1f;
|
||||
float_writer->WriteBatch(1, nullptr, nullptr, &value);
|
||||
}
|
||||
|
||||
// Write the Double column
|
||||
parquet::DoubleWriter* double_writer =
|
||||
static_cast<parquet::DoubleWriter*>(rg_writer->NextColumn());
|
||||
for (int i = 0; i < NUM_ROWS_PER_ROW_GROUP; i++) {
|
||||
double value = i * 1.1111111;
|
||||
double_writer->WriteBatch(1, nullptr, nullptr, &value);
|
||||
}
|
||||
|
||||
// Write the ByteArray column. Make every alternate values NULL
|
||||
parquet::ByteArrayWriter* ba_writer =
|
||||
static_cast<parquet::ByteArrayWriter*>(rg_writer->NextColumn());
|
||||
for (int i = 0; i < NUM_ROWS_PER_ROW_GROUP; i++) {
|
||||
parquet::ByteArray value;
|
||||
char hello[FIXED_LENGTH] = "parquet";
|
||||
hello[7] = static_cast<char>(static_cast<int>('0') + i / 100);
|
||||
hello[8] = static_cast<char>(static_cast<int>('0') + (i / 10) % 10);
|
||||
hello[9] = static_cast<char>(static_cast<int>('0') + i % 10);
|
||||
if (i % 2 == 0) {
|
||||
int16_t definition_level = 1;
|
||||
value.ptr = reinterpret_cast<const uint8_t*>(&hello[0]);
|
||||
value.len = FIXED_LENGTH;
|
||||
ba_writer->WriteBatch(1, &definition_level, nullptr, &value);
|
||||
} else {
|
||||
int16_t definition_level = 0;
|
||||
ba_writer->WriteBatch(1, &definition_level, nullptr, nullptr);
|
||||
}
|
||||
}
|
||||
|
||||
// Write the FixedLengthByteArray column
|
||||
parquet::FixedLenByteArrayWriter* flba_writer =
|
||||
static_cast<parquet::FixedLenByteArrayWriter*>(rg_writer->NextColumn());
|
||||
for (int i = 0; i < NUM_ROWS_PER_ROW_GROUP; i++) {
|
||||
parquet::FixedLenByteArray value;
|
||||
char v = static_cast<char>(i);
|
||||
char flba[FIXED_LENGTH] = {v, v, v, v, v, v, v, v, v, v};
|
||||
value.ptr = reinterpret_cast<const uint8_t*>(&flba[0]);
|
||||
|
||||
flba_writer->WriteBatch(1, nullptr, nullptr, &value);
|
||||
}
|
||||
|
||||
// Close the ParquetFileWriter
|
||||
file_writer->Close();
|
||||
|
||||
// Write the bytes to file
|
||||
DCHECK(out_file->Close().ok());
|
||||
} catch (const std::exception& e) {
|
||||
std::cerr << "Parquet write error: " << e.what() << std::endl;
|
||||
return -1;
|
||||
}
|
||||
|
||||
/**********************************************************************************
|
||||
PARQUET READER EXAMPLE
|
||||
**********************************************************************************/
|
||||
|
||||
try {
|
||||
// Create a ParquetReader instance
|
||||
std::unique_ptr<parquet::ParquetFileReader> parquet_reader =
|
||||
parquet::ParquetFileReader::OpenFile(PARQUET_FILENAME, false);
|
||||
|
||||
// Get the File MetaData
|
||||
std::shared_ptr<parquet::FileMetaData> file_metadata = parquet_reader->metadata();
|
||||
|
||||
// Get the number of RowGroups
|
||||
int num_row_groups = file_metadata->num_row_groups();
|
||||
assert(num_row_groups == 1);
|
||||
|
||||
// Get the number of Columns
|
||||
int num_columns = file_metadata->num_columns();
|
||||
assert(num_columns == 8);
|
||||
|
||||
// Iterate over all the RowGroups in the file
|
||||
for (int r = 0; r < num_row_groups; ++r) {
|
||||
// Get the RowGroup Reader
|
||||
std::shared_ptr<parquet::RowGroupReader> row_group_reader =
|
||||
parquet_reader->RowGroup(r);
|
||||
|
||||
int64_t values_read = 0;
|
||||
int64_t rows_read = 0;
|
||||
int16_t definition_level;
|
||||
int16_t repetition_level;
|
||||
int i;
|
||||
std::shared_ptr<parquet::ColumnReader> column_reader;
|
||||
|
||||
ARROW_UNUSED(rows_read); // prevent warning in release build
|
||||
|
||||
// Get the Column Reader for the boolean column
|
||||
column_reader = row_group_reader->Column(0);
|
||||
parquet::BoolReader* bool_reader =
|
||||
static_cast<parquet::BoolReader*>(column_reader.get());
|
||||
|
||||
// Read all the rows in the column
|
||||
i = 0;
|
||||
while (bool_reader->HasNext()) {
|
||||
bool value;
|
||||
// Read one value at a time. The number of rows read is returned. values_read
|
||||
// contains the number of non-null rows
|
||||
rows_read = bool_reader->ReadBatch(1, nullptr, nullptr, &value, &values_read);
|
||||
// Ensure only one value is read
|
||||
assert(rows_read == 1);
|
||||
// There are no NULL values in the rows written
|
||||
assert(values_read == 1);
|
||||
// Verify the value written
|
||||
bool expected_value = ((i % 2) == 0) ? true : false;
|
||||
assert(value == expected_value);
|
||||
i++;
|
||||
}
|
||||
|
||||
// Get the Column Reader for the Int32 column
|
||||
column_reader = row_group_reader->Column(1);
|
||||
parquet::Int32Reader* int32_reader =
|
||||
static_cast<parquet::Int32Reader*>(column_reader.get());
|
||||
// Read all the rows in the column
|
||||
i = 0;
|
||||
while (int32_reader->HasNext()) {
|
||||
int32_t value;
|
||||
// Read one value at a time. The number of rows read is returned. values_read
|
||||
// contains the number of non-null rows
|
||||
rows_read = int32_reader->ReadBatch(1, nullptr, nullptr, &value, &values_read);
|
||||
// Ensure only one value is read
|
||||
assert(rows_read == 1);
|
||||
// There are no NULL values in the rows written
|
||||
assert(values_read == 1);
|
||||
// Verify the value written
|
||||
assert(value == i);
|
||||
i++;
|
||||
}
|
||||
|
||||
// Get the Column Reader for the Int64 column
|
||||
column_reader = row_group_reader->Column(2);
|
||||
parquet::Int64Reader* int64_reader =
|
||||
static_cast<parquet::Int64Reader*>(column_reader.get());
|
||||
// Read all the rows in the column
|
||||
i = 0;
|
||||
while (int64_reader->HasNext()) {
|
||||
int64_t value;
|
||||
// Read one value at a time. The number of rows read is returned. values_read
|
||||
// contains the number of non-null rows
|
||||
rows_read = int64_reader->ReadBatch(1, &definition_level, &repetition_level,
|
||||
&value, &values_read);
|
||||
// Ensure only one value is read
|
||||
assert(rows_read == 1);
|
||||
// There are no NULL values in the rows written
|
||||
assert(values_read == 1);
|
||||
// Verify the value written
|
||||
int64_t expected_value = i * 1000 * 1000;
|
||||
expected_value *= 1000 * 1000;
|
||||
assert(value == expected_value);
|
||||
if ((i % 2) == 0) {
|
||||
assert(repetition_level == 1);
|
||||
} else {
|
||||
assert(repetition_level == 0);
|
||||
}
|
||||
i++;
|
||||
}
|
||||
|
||||
// Get the Column Reader for the Int96 column
|
||||
column_reader = row_group_reader->Column(3);
|
||||
parquet::Int96Reader* int96_reader =
|
||||
static_cast<parquet::Int96Reader*>(column_reader.get());
|
||||
// Read all the rows in the column
|
||||
i = 0;
|
||||
while (int96_reader->HasNext()) {
|
||||
parquet::Int96 value;
|
||||
// Read one value at a time. The number of rows read is returned. values_read
|
||||
// contains the number of non-null rows
|
||||
rows_read = int96_reader->ReadBatch(1, nullptr, nullptr, &value, &values_read);
|
||||
// Ensure only one value is read
|
||||
assert(rows_read == 1);
|
||||
// There are no NULL values in the rows written
|
||||
assert(values_read == 1);
|
||||
// Verify the value written
|
||||
parquet::Int96 expected_value;
|
||||
ARROW_UNUSED(expected_value); // prevent warning in release build
|
||||
expected_value.value[0] = i;
|
||||
expected_value.value[1] = i + 1;
|
||||
expected_value.value[2] = i + 2;
|
||||
for (int j = 0; j < 3; j++) {
|
||||
assert(value.value[j] == expected_value.value[j]);
|
||||
}
|
||||
i++;
|
||||
}
|
||||
|
||||
// Get the Column Reader for the Float column
|
||||
column_reader = row_group_reader->Column(4);
|
||||
parquet::FloatReader* float_reader =
|
||||
static_cast<parquet::FloatReader*>(column_reader.get());
|
||||
// Read all the rows in the column
|
||||
i = 0;
|
||||
while (float_reader->HasNext()) {
|
||||
float value;
|
||||
// Read one value at a time. The number of rows read is returned. values_read
|
||||
// contains the number of non-null rows
|
||||
rows_read = float_reader->ReadBatch(1, nullptr, nullptr, &value, &values_read);
|
||||
// Ensure only one value is read
|
||||
assert(rows_read == 1);
|
||||
// There are no NULL values in the rows written
|
||||
assert(values_read == 1);
|
||||
// Verify the value written
|
||||
float expected_value = static_cast<float>(i) * 1.1f;
|
||||
assert(value == expected_value);
|
||||
i++;
|
||||
}
|
||||
|
||||
// Get the Column Reader for the Double column
|
||||
column_reader = row_group_reader->Column(5);
|
||||
parquet::DoubleReader* double_reader =
|
||||
static_cast<parquet::DoubleReader*>(column_reader.get());
|
||||
// Read all the rows in the column
|
||||
i = 0;
|
||||
while (double_reader->HasNext()) {
|
||||
double value;
|
||||
// Read one value at a time. The number of rows read is returned. values_read
|
||||
// contains the number of non-null rows
|
||||
rows_read = double_reader->ReadBatch(1, nullptr, nullptr, &value, &values_read);
|
||||
// Ensure only one value is read
|
||||
assert(rows_read == 1);
|
||||
// There are no NULL values in the rows written
|
||||
assert(values_read == 1);
|
||||
// Verify the value written
|
||||
double expected_value = i * 1.1111111;
|
||||
assert(value == expected_value);
|
||||
i++;
|
||||
}
|
||||
|
||||
// Get the Column Reader for the ByteArray column
|
||||
column_reader = row_group_reader->Column(6);
|
||||
parquet::ByteArrayReader* ba_reader =
|
||||
static_cast<parquet::ByteArrayReader*>(column_reader.get());
|
||||
// Read all the rows in the column
|
||||
i = 0;
|
||||
while (ba_reader->HasNext()) {
|
||||
parquet::ByteArray value;
|
||||
// Read one value at a time. The number of rows read is returned. values_read
|
||||
// contains the number of non-null rows
|
||||
rows_read =
|
||||
ba_reader->ReadBatch(1, &definition_level, nullptr, &value, &values_read);
|
||||
// Ensure only one value is read
|
||||
assert(rows_read == 1);
|
||||
// Verify the value written
|
||||
char expected_value[FIXED_LENGTH] = "parquet";
|
||||
ARROW_UNUSED(expected_value); // prevent warning in release build
|
||||
expected_value[7] = static_cast<char>('0' + i / 100);
|
||||
expected_value[8] = static_cast<char>('0' + (i / 10) % 10);
|
||||
expected_value[9] = static_cast<char>('0' + i % 10);
|
||||
if (i % 2 == 0) { // only alternate values exist
|
||||
// There are no NULL values in the rows written
|
||||
assert(values_read == 1);
|
||||
assert(value.len == FIXED_LENGTH);
|
||||
assert(memcmp(value.ptr, &expected_value[0], FIXED_LENGTH) == 0);
|
||||
assert(definition_level == 1);
|
||||
} else {
|
||||
// There are NULL values in the rows written
|
||||
assert(values_read == 0);
|
||||
assert(definition_level == 0);
|
||||
}
|
||||
i++;
|
||||
}
|
||||
|
||||
// Get the Column Reader for the FixedLengthByteArray column
|
||||
column_reader = row_group_reader->Column(7);
|
||||
parquet::FixedLenByteArrayReader* flba_reader =
|
||||
static_cast<parquet::FixedLenByteArrayReader*>(column_reader.get());
|
||||
// Read all the rows in the column
|
||||
i = 0;
|
||||
while (flba_reader->HasNext()) {
|
||||
parquet::FixedLenByteArray value;
|
||||
// Read one value at a time. The number of rows read is returned. values_read
|
||||
// contains the number of non-null rows
|
||||
rows_read = flba_reader->ReadBatch(1, nullptr, nullptr, &value, &values_read);
|
||||
// Ensure only one value is read
|
||||
assert(rows_read == 1);
|
||||
// There are no NULL values in the rows written
|
||||
assert(values_read == 1);
|
||||
// Verify the value written
|
||||
char v = static_cast<char>(i);
|
||||
char expected_value[FIXED_LENGTH] = {v, v, v, v, v, v, v, v, v, v};
|
||||
assert(memcmp(value.ptr, &expected_value[0], FIXED_LENGTH) == 0);
|
||||
i++;
|
||||
}
|
||||
}
|
||||
} catch (const std::exception& e) {
|
||||
std::cerr << "Parquet read error: " << e.what() << std::endl;
|
||||
return -1;
|
||||
}
|
||||
|
||||
std::cout << "Parquet Writing and Reading Complete" << std::endl;
|
||||
|
||||
return 0;
|
||||
}
|
@@ -1,71 +0,0 @@
|
||||
// Licensed to the Apache Software Foundation (ASF) under one
|
||||
// or more contributor license agreements. See the NOTICE file
|
||||
// distributed with this work for additional information
|
||||
// regarding copyright ownership. The ASF licenses this file
|
||||
// to you under the Apache License, Version 2.0 (the
|
||||
// "License"); you may not use this file except in compliance
|
||||
// with the License. You may obtain a copy of the License at
|
||||
//
|
||||
// http://www.apache.org/licenses/LICENSE-2.0
|
||||
//
|
||||
// Unless required by applicable law or agreed to in writing,
|
||||
// software distributed under the License is distributed on an
|
||||
// "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
|
||||
// KIND, either express or implied. See the License for the
|
||||
// specific language governing permissions and limitations
|
||||
// under the License.
|
||||
|
||||
#include <arrow/io/file.h>
|
||||
#include <arrow/util/logging.h>
|
||||
|
||||
#include <parquet/api/reader.h>
|
||||
#include <parquet/api/writer.h>
|
||||
|
||||
using parquet::ConvertedType;
|
||||
using parquet::Repetition;
|
||||
using parquet::Type;
|
||||
using parquet::schema::GroupNode;
|
||||
using parquet::schema::PrimitiveNode;
|
||||
|
||||
constexpr int FIXED_LENGTH = 10;
|
||||
|
||||
static std::shared_ptr<GroupNode> SetupSchema() {
|
||||
parquet::schema::NodeVector fields;
|
||||
// Create a primitive node named 'boolean_field' with type:BOOLEAN,
|
||||
// repetition:REQUIRED
|
||||
fields.push_back(PrimitiveNode::Make("boolean_field", Repetition::REQUIRED,
|
||||
Type::BOOLEAN, ConvertedType::NONE));
|
||||
|
||||
// Create a primitive node named 'int32_field' with type:INT32, repetition:REQUIRED,
|
||||
// logical type:TIME_MILLIS
|
||||
fields.push_back(PrimitiveNode::Make("int32_field", Repetition::REQUIRED, Type::INT32,
|
||||
ConvertedType::TIME_MILLIS));
|
||||
|
||||
// Create a primitive node named 'int64_field' with type:INT64, repetition:REPEATED
|
||||
fields.push_back(PrimitiveNode::Make("int64_field", Repetition::REPEATED, Type::INT64,
|
||||
ConvertedType::NONE));
|
||||
|
||||
fields.push_back(PrimitiveNode::Make("int96_field", Repetition::REQUIRED, Type::INT96,
|
||||
ConvertedType::NONE));
|
||||
|
||||
fields.push_back(PrimitiveNode::Make("float_field", Repetition::REQUIRED, Type::FLOAT,
|
||||
ConvertedType::NONE));
|
||||
|
||||
fields.push_back(PrimitiveNode::Make("double_field", Repetition::REQUIRED, Type::DOUBLE,
|
||||
ConvertedType::NONE));
|
||||
|
||||
// Create a primitive node named 'ba_field' with type:BYTE_ARRAY, repetition:OPTIONAL
|
||||
fields.push_back(PrimitiveNode::Make("ba_field", Repetition::OPTIONAL, Type::BYTE_ARRAY,
|
||||
ConvertedType::NONE));
|
||||
|
||||
// Create a primitive node named 'flba_field' with type:FIXED_LEN_BYTE_ARRAY,
|
||||
// repetition:REQUIRED, field_length = FIXED_LENGTH
|
||||
fields.push_back(PrimitiveNode::Make("flba_field", Repetition::REQUIRED,
|
||||
Type::FIXED_LEN_BYTE_ARRAY, ConvertedType::NONE,
|
||||
FIXED_LENGTH));
|
||||
|
||||
// Create a GroupNode named 'schema' using the primitive nodes defined above
|
||||
// This GroupNode is the root node of the schema tree
|
||||
return std::static_pointer_cast<GroupNode>(
|
||||
GroupNode::Make("schema", Repetition::REQUIRED, fields));
|
||||
}
|
@@ -1,6 +0,0 @@
|
||||
#!/bin/bash
|
||||
|
||||
mkdir src
|
||||
cat deps.log | while read f; do cp --parents $f src/; done;
|
||||
mv src/home/mario/Desktop/arrow/cpp/* src/
|
||||
rm -r src/home/
|
@@ -1,171 +0,0 @@
|
||||
#!/bin/bash
|
||||
|
||||
#-----------------------------------------------------------------------------#
|
||||
|
||||
# NOTE: before starting to extract the minimal required sources and dependencies
|
||||
# run
|
||||
# $ cd cpp/
|
||||
# $ cmake -D ARROW_PARQUET=ON
|
||||
# in the arrow repository
|
||||
|
||||
# provide
|
||||
# - local path of clone of https://github.com/apache/arrow.git
|
||||
# - name/path of main .hpp file of cython extension
|
||||
repo="$1"
|
||||
main="$2"
|
||||
depf="$3"
|
||||
|
||||
# check CLI arguments
|
||||
if [ -z "$repo" ] || [ -z "$main" ] || [ -z "$depf" ]; then
|
||||
echo "please provide..."
|
||||
echo "1. local path of arrow repository"
|
||||
echo "2. name of main .hpp/.cpp"
|
||||
echo "3. desired name of dependency file"
|
||||
echo -e "example:\n./setup-sources.sh /home/mario/Desktop/Record_Evolution/arrow/ reader-writer.cc deps.log"
|
||||
exit 1
|
||||
fi
|
||||
|
||||
echo -e "extracting sources from/for \n1: ${repo}\n2: ${main}\n3: ${depf}\n"
|
||||
|
||||
# make sure the dependency file is empty
|
||||
rm -f ${depf}
|
||||
touch ${depf}
|
||||
|
||||
# define maximal recursion depth
|
||||
maxdep=8
|
||||
|
||||
#-----------------------------------------------------------------------------#
|
||||
|
||||
# define function to list dependencies of source file in repository recursively
|
||||
listDependencies()
|
||||
{
|
||||
rep="$1"
|
||||
src="$2"
|
||||
dep="$3"
|
||||
rec="$4"
|
||||
|
||||
echo -e "\nstarting 'listDependencies()' for\n1. ${rep}\n2. ${src}\n3. ${dep}\n4. ${rec}"
|
||||
|
||||
# generate dependency file (and remove resulting object file)
|
||||
echo -e "g++ -c -MMD ${src} -I ${rep}cpp/src/\n"
|
||||
g++ -c -MMD ${src} -I ${rep}cpp/src/
|
||||
|
||||
# derive name of dependency and object files
|
||||
depf=$(basename ${src} | sed 's/.cc/.d/g')
|
||||
objf=$(basename ${src} | sed 's/.cc/.o/g')
|
||||
rm ${objf}
|
||||
|
||||
# list dependencies by
|
||||
# 1. removing header
|
||||
# 2. remove source itself
|
||||
# 3. delete leading spaces
|
||||
# 4. delete trailing backslashs
|
||||
# 5. remove empty lines
|
||||
cat ${depf} | grep ${rep} | grep -v ${src} | tr -d "^ " | tr -d "\\\\" | awk 'NF' > listdep.log
|
||||
# rm ${depf}
|
||||
|
||||
while IFS= read -r fs
|
||||
do
|
||||
echo "$fs"
|
||||
# check if dependency is already in the list
|
||||
if grep -Fxq "$fs" "$dep"
|
||||
then
|
||||
echo "dep exist"
|
||||
else
|
||||
echo "dep does not exist yet => adding it"
|
||||
# add dependency to list
|
||||
echo "$fs" >> ${dep}
|
||||
# check for corresponding source file
|
||||
fssourc=$(echo ${fs} | sed 's/.h$/.cc/g' | sed 's/.hpp$/.cpp/g')
|
||||
echo ${fssourc}
|
||||
if [ -f "$fssourc" ]
|
||||
then
|
||||
echo "source file exists"
|
||||
# list nested dependencies
|
||||
if [ "$rec" -lt "$maxdep" ]
|
||||
then
|
||||
# increment recursion depth
|
||||
recinc=$(($rec+1))
|
||||
# call recursion
|
||||
listDependencies ${rep} ${fssourc} ${dep} ${recinc}
|
||||
else
|
||||
echo "maximal recursion depth exceeded"
|
||||
fi
|
||||
else
|
||||
echo "source file does not exist"
|
||||
fi
|
||||
fi
|
||||
echo ""
|
||||
done < listdep.log
|
||||
|
||||
# cat listdep.log | while read fs
|
||||
# do
|
||||
# echo $fs
|
||||
# # check if dependency is already in the list
|
||||
# inlist=$(cat listdep.log | grep ${fs} | wc -l)
|
||||
# echo ${inlist}
|
||||
# # check for any corresponding source files
|
||||
# # if [ -f ]
|
||||
# done
|
||||
}
|
||||
|
||||
#-----------------------------------------------------------------------------#
|
||||
|
||||
# call function to list dependencies (recursively)
|
||||
listDependencies ${repo} ${main} ${depf} 0
|
||||
|
||||
# # generate dependency file (and remove resulting object file)
|
||||
# echo -e "generate dependencies:\ng++ -c -MMD ${main} -I ./ -I ${repo}cpp/src/\n"
|
||||
# g++ -c -MMD ${main} -I ${repo}cpp/src/
|
||||
# rm $(echo ${main} | sed 's/.cc/.o/g')
|
||||
#
|
||||
# # derive name of dependency file
|
||||
# dep=$(echo ${main} | sed 's/.cc/.d/g')
|
||||
#
|
||||
# if [ -f "$dep" ]; then
|
||||
#
|
||||
# # list dependencies
|
||||
# cat ${dep} | sed 's/ /\n/g' | awk 'NF' | grep -v '\\' | grep '\/' > deps.log
|
||||
#
|
||||
# # extract list of headers
|
||||
# cat deps.log | sed ':a;N;$!ba;s/\n/ /g' > deps-headers.log
|
||||
# echo "list of required headers ($(cat deps.log | wc -l))"
|
||||
# cat deps-headers.log
|
||||
# echo ""
|
||||
#
|
||||
# # imply list of sources
|
||||
# cat deps.log | sed 's/.h$/.cc/g' | sed 's/.hpp$/.cpp/g' > sources_raw.log
|
||||
# cat sources_raw.log | while read f
|
||||
# do
|
||||
# if [ -f "$f" ]; then
|
||||
# echo $f >> sources_check.log
|
||||
# fi
|
||||
# done
|
||||
# cat sources_check.log | sed ':a;N;$!ba;s/\n/ /g' > deps-sources.log
|
||||
# echo "list of required sources ($(cat sources_check.log | wc -l))"
|
||||
# cat deps-sources.log
|
||||
# echo ""
|
||||
#
|
||||
# # remove all temporary files
|
||||
# rm ${dep} deps.log
|
||||
# rm sources_raw.log sources_check.log
|
||||
#
|
||||
# # copy required headers and sources
|
||||
# echo -e "copy required headers and sources"
|
||||
# mkdir temp/
|
||||
# cp --parents `cat deps-headers.log` temp/
|
||||
# cp --parents `cat deps-sources.log` temp/
|
||||
# mv temp${repo}cpp/src/* ./
|
||||
# rm -r temp
|
||||
#
|
||||
# # remove dependencies
|
||||
# #rm deps-headers.log deps-sources.log
|
||||
#
|
||||
# # show files
|
||||
# ls -lh
|
||||
#
|
||||
# else
|
||||
#
|
||||
# echo -e "\nERROR: failed to generate dependency file\n"
|
||||
#
|
||||
# fi
|
@@ -1,26 +0,0 @@
|
||||
|
||||
FROM ubuntu:19.10
|
||||
|
||||
RUN apt-get update -y && apt-get install -y \
|
||||
apt-utils \
|
||||
git g++ \
|
||||
make cmake \
|
||||
pkg-config \
|
||||
#build-essentials \
|
||||
python3 \
|
||||
python3-setuptools \
|
||||
cython3 \
|
||||
python3-numpy
|
||||
|
||||
RUN git clone https://github.com/apache/arrow.git --single-branch --depth=1
|
||||
|
||||
COPY . ./
|
||||
|
||||
RUN chmod u+x ./build_arrow_cpp.sh
|
||||
RUN chmod u+x ./build_arrow_python.sh
|
||||
RUN ./build_arrow_cpp.sh
|
||||
RUN ./build_arrow_python.sh
|
||||
|
||||
#RUN chmod u+x ./build_arrow.sh
|
||||
#CMD ["./build_arrow.sh"]
|
||||
CMD ["sleep 1d"]
|
@@ -1,5 +0,0 @@
|
||||
#!/bin/bash
|
||||
|
||||
docker build . --tag=pyarrowbuild:latest
|
||||
|
||||
docker run -it pyarrowbuild:latest /bin/bash
|
@@ -1,65 +0,0 @@
|
||||
#!/bin/bash
|
||||
|
||||
sleep infinity
|
||||
|
||||
startts=$(date)
|
||||
echo "starting build process at ${startts}..."
|
||||
|
||||
echo -e "\nhome directory is..."
|
||||
pwd
|
||||
|
||||
echo -e "\ncloning apache/arrow..."
|
||||
git clone https://github.com/apache/arrow.git --single-branch --depth=1
|
||||
|
||||
echo -e "\nls -lh /\n"
|
||||
ls -lh /
|
||||
|
||||
echo -e "\nls -lh arrow/\n"
|
||||
ls -lh arrow/
|
||||
|
||||
echo -e "\nls -lh arrow/python/\n"
|
||||
ls -lh arrow/python
|
||||
|
||||
|
||||
mkdir arrow/cpp/build
|
||||
pushd arrow/cpp/build
|
||||
|
||||
cmake -DCMAKE_INSTALL_PREFIX=$ARROW_HOME \
|
||||
-DCMAKE_INSTALL_LIBDIR=lib \
|
||||
-DARROW_WITH_BZ2=ON \
|
||||
-DARROW_WITH_ZLIB=ON \
|
||||
-DARROW_WITH_ZSTD=ON \
|
||||
-DARROW_WITH_LZ4=ON \
|
||||
-DARROW_WITH_SNAPPY=ON \
|
||||
-DARROW_WITH_BROTLI=ON \
|
||||
-DARROW_PARQUET=ON \
|
||||
-DARROW_PYTHON=ON \
|
||||
-DARROW_BUILD_TESTS=OFF \
|
||||
-DARROW_WITH_HDFS=OFF \
|
||||
..
|
||||
|
||||
make -j4
|
||||
make install
|
||||
popd
|
||||
|
||||
#cython --version
|
||||
cython3 --version
|
||||
|
||||
pushd arrow/python
|
||||
export ARROW_LIB_DIR=/lib/
|
||||
export PYARROW_WITH_PARQUET=1
|
||||
export PYARROW_WITH_CUDA=0
|
||||
export PYARROW_WITH_FlIGHT=0
|
||||
export PYARROW_WITH_DATASET=0
|
||||
export PYARROW_WITH_ORC=0
|
||||
export PYARROW_WITH_PLASMA=0
|
||||
export PYARROW_WITH_S3FS=0
|
||||
export PYARROW_WITH_HDFS=0
|
||||
export PYARROW_WITH_GANDIVA=0
|
||||
python3 setup.py build_ext --inplace
|
||||
popd
|
||||
|
||||
echo " started build process at ${startts} ..."
|
||||
finishts=$(date)
|
||||
echo "finishing build process at ${finishts}..."
|
||||
|
@@ -1,23 +0,0 @@
|
||||
#!/bin/bash
|
||||
|
||||
mkdir arrow/cpp/build
|
||||
pushd arrow/cpp/build
|
||||
|
||||
cmake -DCMAKE_INSTALL_PREFIX=$ARROW_HOME \
|
||||
-DCMAKE_INSTALL_LIBDIR=lib \
|
||||
-DARROW_WITH_BZ2=ON \
|
||||
-DARROW_WITH_ZLIB=ON \
|
||||
-DARROW_WITH_ZSTD=ON \
|
||||
-DARROW_WITH_LZ4=ON \
|
||||
-DARROW_WITH_SNAPPY=ON \
|
||||
-DARROW_WITH_BROTLI=ON \
|
||||
-DARROW_PARQUET=ON \
|
||||
-DARROW_PYTHON=ON \
|
||||
-DARROW_BUILD_TESTS=OFF \
|
||||
-DARROW_WITH_HDFS=OFF \
|
||||
-DARROW_WITH_IPC=OFF \
|
||||
..
|
||||
|
||||
make -j4
|
||||
make install
|
||||
popd
|
@@ -1,15 +0,0 @@
|
||||
#!/bin/bash
|
||||
|
||||
pushd arrow/python
|
||||
export PYARROW_WITH_PARQUET=1
|
||||
export PYARROW_WITH_CUDA=0
|
||||
export PYARROW_WITH_FlIGHT=0
|
||||
export PYARROW_WITH_DATASET=0
|
||||
export PYARROW_WITH_ORC=0
|
||||
export PYARROW_WITH_PLASMA=0
|
||||
export PYARROW_WITH_S3FS=0
|
||||
export PYARROW_WITH_HDFS=0
|
||||
export PYARROW_WITH_GANDIVA=0
|
||||
# python3 setup.py build_ext --inplace
|
||||
python3 setup.py install
|
||||
popd
|
@@ -1,23 +0,0 @@
|
||||
|
||||
build :
|
||||
docker build . --tag pyarrowbuild
|
||||
|
||||
run :
|
||||
docker run -it pyarrowbuild:latest
|
||||
|
||||
run-bash :
|
||||
docker run -it --volume=$(pwd)/build:/home pyarrowbuild:latest /bin/bash
|
||||
|
||||
run-volume :
|
||||
docker run -it -v /home/pirate/pyarrow/build/:/arrow/python/ pyarrowbuild:latest
|
||||
#sudo docker run -it --volume=$(pwd)/build:/home ubuntu:latest /bin/bash
|
||||
|
||||
rm-container :
|
||||
cont=$(docker ps -a | tail -n 26 | awk '{print $NF}' | sed ':a;N;$!ba;s/\n/ /g')
|
||||
echo ${cont}
|
||||
docker rm ${cont}
|
||||
|
||||
rm-image :
|
||||
img=$(docker image ls --quiet | sed ':a;N;$!ba;s/\n/ /g')
|
||||
docker image rm ${img}
|
||||
|
@@ -1,18 +0,0 @@
|
||||
|
||||
import pyarrow.parquet as pq
|
||||
import pyarrow.csv as pv
|
||||
|
||||
csvfile = 'pressureVacuum.csv'
|
||||
|
||||
tb = pv.read_csv(csvfile,parse_options=pv.ParseOptions(delimiter=','))
|
||||
|
||||
print(tb)
|
||||
|
||||
parquetfile = 'pressureVacuum.parquet'
|
||||
|
||||
pq.write_table(tb,parquetfile,compression='BROTLI')
|
||||
# {'NONE', 'SNAPPY', 'GZIP', 'LZO', 'BROTLI', 'LZ4', 'ZSTD'}
|
||||
|
||||
df = pq.read_table(parquetfile,columns=None)
|
||||
|
||||
print(df)
|
@@ -1,8 +0,0 @@
|
||||
#!/bin/bash
|
||||
|
||||
if [ -z "$1" ]
|
||||
then
|
||||
exit 1
|
||||
fi
|
||||
|
||||
scp $1 pirate@mf-pi-40:/home/pirate/pyarrow/
|
@@ -1,4 +0,0 @@
|
||||
[build-system]
|
||||
requires = [
|
||||
"setuptools"
|
||||
]
|
29
pip/makefile
29
pip/makefile
@@ -1,29 +0,0 @@
|
||||
# --------------------------------------------------------------------------- #
|
||||
|
||||
SHELL := /bin/bash
|
||||
|
||||
publish: sdist upload
|
||||
|
||||
sdist: ../cython/py_imc_termite.pyx ../cython/imc_termite.pxd ../cython/py_imc_termite.cpp
|
||||
cp -v $? ./
|
||||
cp -v $(shell ls ../lib/imc_*.hpp) ./
|
||||
tail -n 212 ../README.md > ./README.md
|
||||
cp -v ../LICENSE ./
|
||||
python3 setup.py sdist
|
||||
|
||||
# authentication:
|
||||
# - username: __token__
|
||||
# - password: <token value including pypi-prefix>
|
||||
upload:
|
||||
python3 -m twine upload dist/$(shell ls -t dist/ | head -n1)
|
||||
|
||||
clean:
|
||||
rm -rvf dist/
|
||||
rm -rvf *.egg-info
|
||||
rm -rvf build/
|
||||
rm -rvf cython/
|
||||
rm -vf *.pyx *.pxd
|
||||
rm -vf *.cpp *.c *.hpp
|
||||
rm -vf README.md LICENSE
|
||||
|
||||
# --------------------------------------------------------------------------- #
|
36
pip/setup.py
36
pip/setup.py
@@ -1,36 +0,0 @@
|
||||
|
||||
from setuptools import setup, Extension
|
||||
|
||||
with open("README.md", "r", encoding="utf-8") as fh:
|
||||
long_description = fh.read()
|
||||
|
||||
setup(
|
||||
name="IMCtermite",
|
||||
version="1.2.4",
|
||||
author="Record Evolution GmbH",
|
||||
author_email="mario.fink@record-evolution.de",
|
||||
maintainer="Record Evolution GmbH",
|
||||
license="MIT",
|
||||
description="Enables extraction of measurement data from binary files with extension 'raw' used by proprietary software imcFAMOS/imcSTUDIO and facilitates its storage in open source file formats",
|
||||
keywords="IMC raw imcFAMOS imcSTUDIO imcCRONOS",
|
||||
long_description=long_description,
|
||||
long_description_content_type="text/markdown",
|
||||
url="https://github.com/RecordEvolution/IMCtermite.git",
|
||||
project_urls={
|
||||
"Bug Tracker": "https://github.com/RecordEvolution/IMCtermite/issues",
|
||||
},
|
||||
classifiers=[
|
||||
"Programming Language :: Python :: 3",
|
||||
"License :: OSI Approved :: MIT License",
|
||||
"Operating System :: OS Independent",
|
||||
],
|
||||
ext_modules=[Extension("imc_termite",
|
||||
["py_imc_termite.cpp"],
|
||||
# libraries_dirs=["cython/"],
|
||||
# include_dirs=["3rdparty/pugixml/","lib/"],
|
||||
# depends=["../lib/tdm_termite.hpp"]
|
||||
language='c++',
|
||||
extra_compile_args=['-std=c++17','-Wno-unused-variable'],
|
||||
extra_link_args=['-std=c++17'],
|
||||
)],
|
||||
)
|
@@ -1,19 +1,23 @@
|
||||
# cython: language_level = 3
|
||||
|
||||
# use some C++ STL libraries
|
||||
from libcpp.string cimport string
|
||||
from libcpp.vector cimport vector
|
||||
from libcpp cimport bool
|
||||
|
||||
cdef extern from "imc_raw.hpp" namespace "imc":
|
||||
cdef cppclass imc_termite "imc::raw":
|
||||
cdef extern from "lib/imc_raw.hpp" namespace "imc":
|
||||
|
||||
cdef cppclass cppimctermite "imc::raw":
|
||||
|
||||
# constructor(s)
|
||||
imc_termite() except +
|
||||
imc_termite(string rawfile) except +
|
||||
cppimctermite() except +
|
||||
cppimctermite(string rawfile) except +
|
||||
|
||||
# provide raw file
|
||||
void set_file(string rawfile) except +
|
||||
|
||||
# get JSON list of channels
|
||||
vector[string] get_channels(bool json, bool data) except +
|
||||
|
||||
# print single channel/all channels
|
||||
void print_channel(string channeluuid, string outputdir, char delimiter) except +
|
||||
void print_channels(string outputdir, char delimiter) except +
|
@@ -1,39 +1,39 @@
|
||||
# distutils: language = c++
|
||||
# cython: language_level = 3
|
||||
|
||||
from imc_termite cimport imc_termite
|
||||
from IMCtermite cimport cppimctermite
|
||||
|
||||
import json as jn
|
||||
import decimal
|
||||
# import numpy as np
|
||||
|
||||
cdef class imctermite:
|
||||
|
||||
# C++ instance of class => stack allocated (requires nullary constructor!)
|
||||
cdef imc_termite cpp_imc
|
||||
cdef cppimctermite cppimc
|
||||
|
||||
# constructor
|
||||
def __cinit__(self, string rawfile):
|
||||
self.cpp_imc = imc_termite(rawfile)
|
||||
self.cppimc = cppimctermite(rawfile)
|
||||
|
||||
# provide raw file
|
||||
def submit_file(self,string rawfile):
|
||||
self.cpp_imc.set_file(rawfile)
|
||||
self.cppimc.set_file(rawfile)
|
||||
|
||||
# get JSON list of channels
|
||||
def get_channels(self, bool data):
|
||||
chnlst = self.cpp_imc.get_channels(True,data)
|
||||
def get_channels(self, bool include_data):
|
||||
chnlst = self.cppimc.get_channels(True,include_data)
|
||||
chnlstjn = [jn.loads(chn.decode(errors="ignore")) for chn in chnlst]
|
||||
return chnlstjn
|
||||
|
||||
# print single channel/all channels
|
||||
def print_channel(self, string channeluuid, string outputfile, char delimiter):
|
||||
self.cpp_imc.print_channel(channeluuid,outputfile,delimiter)
|
||||
self.cppimc.print_channel(channeluuid,outputfile,delimiter)
|
||||
def print_channels(self, string outputdir, char delimiter):
|
||||
self.cpp_imc.print_channels(outputdir,delimiter)
|
||||
self.cppimc.print_channels(outputdir,delimiter)
|
||||
|
||||
# print table including channels
|
||||
def print_table(self, string outputfile):
|
||||
chnlst = self.cpp_imc.get_channels(True,True)
|
||||
chnlst = self.cppimc.get_channels(True,True)
|
||||
chnlstjn = [jn.loads(chn.decode(errors="ignore")) for chn in chnlst]
|
||||
with open(outputfile.decode(),'w') as fout:
|
||||
for chn in chnlstjn:
|
@@ -1,4 +1,5 @@
|
||||
include *.hpp
|
||||
include lib/*.hpp
|
||||
include *.cpp
|
||||
include *.pyx
|
||||
include *.pxd
|
||||
include VERSION
|
1
python/VERSION
Normal file
1
python/VERSION
Normal file
@@ -0,0 +1 @@
|
||||
2.0.9
|
@@ -1,137 +0,0 @@
|
||||
|
||||
#-----------------------------------------------------------------------------#
|
||||
|
||||
import raw_eater
|
||||
import raw_meat
|
||||
import pyarrow as pa
|
||||
import pyarrow.parquet as pq
|
||||
from pathlib import Path
|
||||
|
||||
fileobj1 = Path("samples/datasetA/").rglob("*.raw")
|
||||
rawlist1 = [str(fl) for fl in fileobj1]
|
||||
|
||||
fileobj2 = Path("samples/datasetB/").rglob("*.raw")
|
||||
rawlist2 = [str(fl) for fl in fileobj2]
|
||||
|
||||
rawlist = rawlist1 #[rawlist1[0],rawlist1[4],rawlist2[0],rawlist2[6]]
|
||||
for fil in rawlist2 :
|
||||
rawlist.append(fil)
|
||||
rawlist.append("./README.md")
|
||||
|
||||
print("")
|
||||
print(rawlist)
|
||||
print()
|
||||
|
||||
#-----------------------------------------------------------------------------#
|
||||
|
||||
# alternatively create "empty" instance of "raw_eater" and set file names
|
||||
eatraw = raw_eater.raweater()
|
||||
# eatraw.set_file("../smp/pressure_Vacuum.raw".encode())
|
||||
|
||||
# convert every single listed file
|
||||
for rf in rawlist :
|
||||
|
||||
print("converting " + str(rf) + "...\n" + 90*("-") + "\n")
|
||||
|
||||
# setup instance of "raw_eater" and trigger conversion
|
||||
# eatraw = raw_eater.raweater(rf.encode())
|
||||
# eatraw = raw_meat.rawmerger(rf.encode())
|
||||
|
||||
# use global instance of "raw_eater" to set file and perform decoding
|
||||
eatraw.set_file(rf.encode())
|
||||
try :
|
||||
eatraw.do_conversion()
|
||||
except RuntimeError as e :
|
||||
print("conversion failed: " + str(e))
|
||||
|
||||
# check validity of file format
|
||||
if eatraw.validity() :
|
||||
|
||||
# show channel name and its unit
|
||||
entity = eatraw.channel_name().decode(encoding='UTF-8',errors='ignore')
|
||||
unit = eatraw.unit().decode(encoding='UTF-8',errors='ignore')
|
||||
print("\nentity: " + str(entity))
|
||||
print("unit: " + str(unit) + "\n")
|
||||
|
||||
# obtain extracted data
|
||||
xt = eatraw.get_time()
|
||||
yt = eatraw.get_channel()
|
||||
|
||||
# show excerpt of data
|
||||
print("time (length: " + str(len(xt)) + ") \n"
|
||||
+ str(xt[:10]) + "\n...\n" + str(xt[-10:]) + "\n")
|
||||
yttrunc = [round(y,4) for y in yt]
|
||||
print(str(entity) + " (length: " + str(len(yttrunc)) + ") \n"
|
||||
+ str(yttrunc[:10]) + "\n...\n" + str(yttrunc[-10:]) + "\n")
|
||||
|
||||
outname = rf.split('/')[-1].replace('raw','csv')
|
||||
|
||||
print("write output to : " + outname)
|
||||
eatraw.write_table(("output/"+outname).encode(),ord(' '))
|
||||
|
||||
else :
|
||||
|
||||
print("\nerror: invalid/corrupt .raw file")
|
||||
|
||||
print("\n")
|
||||
|
||||
#-----------------------------------------------------------------------------#
|
||||
|
||||
print("convert and merge channels " + "\n" + 90*("-") + "\n")
|
||||
|
||||
# setup new instance to merge channels
|
||||
eatmea = raw_meat.rawmerger(''.encode()) #rawlist[0].encode())
|
||||
|
||||
# add every single channel/file in list
|
||||
for rf in rawlist :
|
||||
print("\nadding channel " + str(rf))
|
||||
try :
|
||||
succ = eatmea.add_channel(rf.encode())
|
||||
print("\nrecent time series: length: " + str(len(eatmea.get_time_series())) + "\n")
|
||||
except RuntimeError as e :
|
||||
print("failed to add channel: " + str(e))
|
||||
|
||||
# show summary of successfully merged channels
|
||||
print("\nmerged channels:\n")
|
||||
|
||||
# write merged table to .csv output
|
||||
eatmea.write_table_all('output/allchannels.csv'.encode(),ord(','))
|
||||
|
||||
# get number of successfully merged channels and their names (+units)
|
||||
numch = eatmea.get_num_channels()
|
||||
chnames = [chnm.decode(encoding='UTF-8',errors='ignore') for chnm in eatmea.get_channel_names()]
|
||||
print("number of channels: " + str(numch))
|
||||
print("channel names: " + str(chnames))
|
||||
|
||||
# obtain final time series
|
||||
timse = eatmea.get_time_series()
|
||||
print("\nfinal time series:\nlength:" + str(len(timse)) + "\n")
|
||||
|
||||
# get time unit and prepend column name
|
||||
chnames.insert(0,"Time ["+str(eatmea.time_unit().decode(encoding='UTF-8',errors='ignore'))+"]")
|
||||
|
||||
# prepare list of pyarrow arrays
|
||||
pyarrs = []
|
||||
pyarrs.append(pa.array(timse))
|
||||
|
||||
for i in range(0,numch) :
|
||||
print("\n" + str(i) + " " + str(chnames[i]))
|
||||
dat = eatmea.get_channel_by_index(i)
|
||||
print("length: " + str(len(dat)))
|
||||
pyarrs.append(pa.array(dat))
|
||||
print("")
|
||||
# print("\npyarrow arrays\n" + str(pyarrs))
|
||||
|
||||
# create pyarrow table from data
|
||||
pyarwtab = pa.Table.from_arrays(pyarrs,chnames)
|
||||
print("\n" + 60*"-" + "\n" + str(pyarwtab) + "\n")
|
||||
|
||||
# write pyarrow table to .parquet file with compression
|
||||
pq.write_table(pyarwtab,'output/allchannels.parquet',compression='BROTLI') # compression='BROTLI', 'SNAPPY')
|
||||
|
||||
# try to read and decode the .parquet file
|
||||
df = pq.read_table('output/allchannels.parquet')
|
||||
print(df.to_pandas())
|
||||
# df.to_pandas().to_csv('allchannels.csv',index=False,encoding='utf-8',sep=",")
|
||||
|
||||
#-----------------------------------------------------------------------------#
|
43
python/examples/multichannel.py
Normal file
43
python/examples/multichannel.py
Normal file
@@ -0,0 +1,43 @@
|
||||
|
||||
import IMCtermite
|
||||
import pandas
|
||||
import datetime
|
||||
|
||||
def add_trigger_time(trigger_time, add_time) :
|
||||
trgts = datetime.datetime.strptime(trigger_time,'%Y-%m-%dT%H:%M:%S')
|
||||
dt = datetime.timedelta(seconds=add_time)
|
||||
return (trgts + dt).strftime('%Y-%m-%dT%H:%M:%S:%f')
|
||||
|
||||
if __name__ == "__main__" :
|
||||
|
||||
# read file and extract data
|
||||
imctm = IMCtermite.imctermite(b"Measurement.raw")
|
||||
chns = imctm.get_channels(True)
|
||||
|
||||
# prepare abscissa
|
||||
xcol = "time ["+chns[0]['xunit']+"]"
|
||||
#xcol = "timestamp"
|
||||
xsts = [add_trigger_time(chns[0]['trigger-time'],tm) for tm in chns[0]['xdata']]
|
||||
|
||||
# sort channels
|
||||
chnnms = sorted([chn['name'] for chn in chns], reverse=False)
|
||||
chnsdict = {}
|
||||
for chn in chns :
|
||||
chnsdict[chn['name']] = chn
|
||||
|
||||
# construct dataframe
|
||||
df = pandas.DataFrame()
|
||||
df[xcol] = pandas.Series(chns[0]['xdata'])
|
||||
#df[xcol] = pandas.Series(xsts)
|
||||
#for idx,chn in enumerate(chns) :
|
||||
for chnnm in chnnms :
|
||||
chn = chnsdict[chnnm]
|
||||
#xcol = (chn['xname'] if chn['xname'] != '' else "x_"+str(idx))+" ["+chn['xunit']+"]"
|
||||
#df[xcol] = pandas.Series(chn['xdata'])
|
||||
ycol = chn['yname']+" ["+chn['yunit']+"]"
|
||||
df[ycol] = pandas.Series(chn['ydata'])
|
||||
|
||||
# show entire dataframe and write file
|
||||
print(df)
|
||||
df.to_csv("Measurement.csv",header=True,sep='\t',index=False)
|
||||
|
@@ -1,11 +1,11 @@
|
||||
|
||||
import imc_termite
|
||||
import IMCtermite
|
||||
import json
|
||||
import os
|
||||
|
||||
# declare and initialize instance of "imctermite" by passing a raw-file
|
||||
try :
|
||||
imcraw = imc_termite.imctermite(b"samples/exampleB.raw")
|
||||
imcraw = IMCtermite.imctermite(b"samples/exampleB.raw")
|
||||
except RuntimeError as e :
|
||||
raise Exception("failed to load/parse raw-file: " + str(e))
|
||||
|
||||
@@ -21,18 +21,18 @@ if len(channelsdata) > 0 :
|
||||
|
||||
print(len(chnydata))
|
||||
print(len(chnxdata))
|
||||
print()
|
||||
|
||||
# print the channels into a specific directory
|
||||
imcraw.print_channels(b"./data",ord(','))
|
||||
imcraw.print_channels(b"/tmp/",ord(','))
|
||||
|
||||
# print all channels separately
|
||||
idx = 0
|
||||
for chn in channels :
|
||||
print(str(idx)+" : "+chn['name']+" : "+chn['uuid'])
|
||||
filname = os.path.join("./data",str(idx) + "_" + chn['name']+".csv")
|
||||
for i,chn in enumerate(channels) :
|
||||
print(str(i)+" : "+chn['name']+" : "+chn['uuid'])
|
||||
filname = os.path.join("/tmp/",str(i) + "_" + chn['name']+".csv")
|
||||
print(filname)
|
||||
imcraw.print_channel(chn['uuid'].encode(),filname.encode(),ord(','))
|
||||
idx = idx + 1
|
||||
|
||||
# print all channels in single file
|
||||
# imcraw.print_table(b"./data/allchannels.csv")
|
||||
imcraw.print_table(b"/tmp/allchannels.csv")
|
||||
|
@@ -1,5 +1,5 @@
|
||||
|
||||
import imc_termite
|
||||
import IMCtermite
|
||||
import json
|
||||
import os
|
||||
|
||||
@@ -15,7 +15,7 @@ for fl in rawlist1:
|
||||
|
||||
# declare and initialize instance of "imctermite" by passing a raw-file
|
||||
try :
|
||||
imcraw = imc_termite.imctermite(fl.encode())
|
||||
imcraw = IMCtermite.imctermite(fl.encode())
|
||||
except RuntimeError as e :
|
||||
raise Exception("failed to load/parse raw-file: " + str(e))
|
||||
|
||||
@@ -24,7 +24,7 @@ for fl in rawlist1:
|
||||
print(json.dumps(channels,indent=4, sort_keys=False))
|
||||
|
||||
# print the channels into a specific directory
|
||||
imcraw.print_channels(b"./")
|
||||
imcraw.print_channels(b"./",ord(','))
|
||||
|
||||
# print all channels in single file
|
||||
imcraw.print_table(("./"+str(os.path.basename(fl).split('.')[0])+"_allchannels.csv").encode())
|
50
python/examples/usage_ext.py
Normal file
50
python/examples/usage_ext.py
Normal file
@@ -0,0 +1,50 @@
|
||||
|
||||
import IMCtermite
|
||||
import json
|
||||
import os
|
||||
import datetime
|
||||
|
||||
# declare and initialize instance of "imctermite" by passing a raw-file
|
||||
try :
|
||||
imcraw = IMCtermite.imctermite(b"samples/sampleB.raw")
|
||||
except RuntimeError as e :
|
||||
raise Exception("failed to load/parse raw-file: " + str(e))
|
||||
|
||||
# obtain list of channels as list of dictionaries (without data)
|
||||
channels = imcraw.get_channels(False)
|
||||
print(json.dumps(channels,indent=4, sort_keys=False))
|
||||
|
||||
# obtain all channels (including full data)
|
||||
channelsdata = imcraw.get_channels(True)
|
||||
|
||||
# everything that follows is an example that specifically makes use only of
|
||||
# the first (index = 0) channel ...
|
||||
idx = 0
|
||||
|
||||
if len(channelsdata) > 0 :
|
||||
|
||||
# get first channel's data
|
||||
chnydata = channelsdata[idx]['ydata']
|
||||
chnxdata = channelsdata[idx]['xdata']
|
||||
print("xdata: " + str(len(chnxdata)))
|
||||
print("ydata: " + str(len(chnydata)))
|
||||
|
||||
# extract trigger-time
|
||||
trigtim = datetime.datetime.fromisoformat(channels[idx]["trigger-time"])
|
||||
print(trigtim)
|
||||
|
||||
# file output of data with absolute timestamp in 1st column
|
||||
filname = os.path.join("./",channelsdata[idx]['name']+".csv")
|
||||
print("writing output into " + filname)
|
||||
with open(filname,'w') as fout :
|
||||
# include column header
|
||||
fout.write( str(channelsdata[idx]['xname']) + '[' + str(channelsdata[idx]['xunit']) + "]"
|
||||
+ ","
|
||||
+ str(channelsdata[idx]['yname']) + '[' + str(channelsdata[idx]['yunit']) + "]"
|
||||
+ "\n" )
|
||||
# add data (introduce time shift according to trigger-time)
|
||||
for row in range(0,len(chnxdata)) :
|
||||
fout.write( str( (trigtim + datetime.timedelta(seconds=chnxdata[row])).isoformat() )
|
||||
+ ","
|
||||
+ str( chnydata[row])
|
||||
+ "\n" )
|
29
python/examples/usage_files.py
Normal file
29
python/examples/usage_files.py
Normal file
@@ -0,0 +1,29 @@
|
||||
|
||||
from IMCtermite import imctermite
|
||||
|
||||
def show_results(imcraw) :
|
||||
|
||||
channels = imcraw.get_channels(False)
|
||||
print(channels)
|
||||
|
||||
channelsData = imcraw.get_channels(True)
|
||||
print("number of channels: " + str(len(channelsData)))
|
||||
|
||||
for (i,chn) in enumerate(channelsData) :
|
||||
print(str(i) + " | " + chn['name'])
|
||||
print(chn['xname'] + " | " + chn['xunit'])
|
||||
print(chn['xdata'][:10])
|
||||
print(chn['yname'] + " | " + chn['yunit'])
|
||||
print(chn['ydata'][:10])
|
||||
|
||||
print("")
|
||||
|
||||
# create instance of 'imctermite'
|
||||
imcraw = imctermite(b'samples/sampleA.raw')
|
||||
|
||||
show_results(imcraw)
|
||||
|
||||
# use previous instance of 'imctermite' to provide new file
|
||||
imcraw.submit_file(b'samples/sampleB.raw')
|
||||
|
||||
show_results(imcraw)
|
45
python/makefile
Normal file
45
python/makefile
Normal file
@@ -0,0 +1,45 @@
|
||||
|
||||
setup:
|
||||
cat ../README.md | grep '^# IMCtermite' -A 50000 > ./README.md
|
||||
#pandoc -f markdown -t rst -o README.rst README.md
|
||||
#python -m rstvalidator README.rst
|
||||
cp -r ../lib ./
|
||||
cp -v ../LICENSE ./
|
||||
|
||||
setup-clean:
|
||||
rm -vf README.md README.rst LICENSE
|
||||
rm -rf lib/
|
||||
|
||||
build: setup
|
||||
python setup.py build
|
||||
|
||||
build-inplace: setup
|
||||
python setup.py build_ext --inplace
|
||||
|
||||
build-sdist: setup
|
||||
python setup.py sdist
|
||||
python -m twine check dist/*
|
||||
|
||||
build-bdist: setup
|
||||
python setup.py bdist
|
||||
python -m twine check dist/*
|
||||
|
||||
build-clean:
|
||||
python setup.py clean --all
|
||||
rm -vf imctermite*.so imctermite*.cpp
|
||||
rm -vf IMCtermite*.so IMCtermite*.cpp
|
||||
rm -rvf dist/ IMCtermite.egg-info/
|
||||
|
||||
cibuildwheel-build: setup
|
||||
cibuildwheel --platform linux
|
||||
|
||||
cibuildwheel-clean:
|
||||
rm -rvf wheelhouse/
|
||||
|
||||
pypi-upload:
|
||||
python -m twine upload dist/$(shell ls -t dist/ | head -n1)
|
||||
|
||||
clean: setup build-clean cibuildwheel-clean setup-clean
|
||||
|
||||
run-example:
|
||||
PYTHONPATH=$(pwd) python examples/usage_files.py
|
@@ -1,24 +0,0 @@
|
||||
|
||||
import pyarrow as pa
|
||||
import numpy as np
|
||||
import pyarrow.parquet as pq
|
||||
|
||||
db = pa.array(np.linspace(10,50,6))
|
||||
print(db)
|
||||
da = pa.array(np.linspace(0,5,6))
|
||||
print(db)
|
||||
|
||||
filenam = 'pyarrow_testtab.parquet'
|
||||
|
||||
patab = pa.Table.from_arrays([da,db],['entity A [unitA]','entity B [unitB]'])
|
||||
print(patab)
|
||||
|
||||
# pq.write_table(patab,filenam,compression='BROTLI')
|
||||
pq.write_table(patab,filenam,compression='SNAPPY')
|
||||
|
||||
df = pq.read_table(filenam)
|
||||
print(df)
|
||||
print(df.to_pandas())
|
||||
|
||||
#import readline
|
||||
#readline.write_history_file('generate_pyarrow_table_and_write_parquet.py')
|
6
python/pyproject.toml
Normal file
6
python/pyproject.toml
Normal file
@@ -0,0 +1,6 @@
|
||||
[build-system]
|
||||
requires = ["setuptools", "wheel","Cython"]
|
||||
build-backend = "setuptools.build_meta"
|
||||
|
||||
[tool.cibuildwheel]
|
||||
before-all = ""
|
23
python/setup.cfg
Normal file
23
python/setup.cfg
Normal file
@@ -0,0 +1,23 @@
|
||||
|
||||
[metadata]
|
||||
name = IMCtermite
|
||||
description = Enables extraction of measurement data from binary files with extension 'raw' used by proprietary software imcFAMOS and imcSTUDIO and facilitates its storage in open source file formats
|
||||
long_description = file: README.md
|
||||
# long_description_content_type = text/x-rst
|
||||
long_description_content_type = text/markdown
|
||||
version = file: VERSION
|
||||
author = Record Evolution GmbH
|
||||
author_email = mario.fink@record-evolution.de
|
||||
maintainer = Record Evolution GmbH
|
||||
url= https://github.com/RecordEvolution/IMCtermite.git
|
||||
license = MIT License
|
||||
license_files = LICENSE
|
||||
keywords = IMC, raw, imcFAMOS, imcSTUDIO, imcCRONOS
|
||||
classifiers =
|
||||
Programming Language :: Python :: 3
|
||||
License :: OSI Approved :: MIT License
|
||||
Operating System :: OS Independent
|
||||
Topic :: Scientific/Engineering
|
||||
Topic :: Software Development :: Libraries :: Python Modules
|
||||
|
||||
[options]
|
21
python/setup.py
Normal file
21
python/setup.py
Normal file
@@ -0,0 +1,21 @@
|
||||
from setuptools import Extension, setup
|
||||
from Cython.Build import cythonize
|
||||
import sys
|
||||
|
||||
print("building on platform: "+sys.platform)
|
||||
|
||||
cmpArgs = {
|
||||
"linux": ['-std=c++17','-Wno-unused-variable'],
|
||||
"darwin": ['-std=c++17','-Wno-unused-variable'],
|
||||
"win32": ['/EHsc','/std:c++17']
|
||||
}
|
||||
|
||||
extension = Extension(
|
||||
"IMCtermite",
|
||||
sources=["IMCtermite.pyx"],
|
||||
extra_compile_args=cmpArgs[sys.platform]
|
||||
)
|
||||
|
||||
setup(
|
||||
ext_modules=cythonize(extension,language_level=3)
|
||||
)
|
Binary file not shown.
@@ -1,12 +1,50 @@
|
||||
##!/bin/bash/
|
||||
|
||||
dir=$1
|
||||
fildir=$1
|
||||
|
||||
#ls ${dir} | while read fn; do echo $fn; cat ${dir}$fn | grep -a "|[A-Z][A-Z]," -o | wc -l; done;
|
||||
if [ -z "${fildir}" ]; then
|
||||
echo "CLI argument missing: provide file or directory" >&2
|
||||
exit 1
|
||||
fi
|
||||
|
||||
#ls ${dir} | while read fn; do echo $fn; cat ${dir}$fn | grep -a "|[A-Z][A-Z]," -o; done;
|
||||
list-markers()
|
||||
{
|
||||
arg="$1"
|
||||
if [ -z "${arg}" ]; then
|
||||
echo "list-markers: missing file argument" >&2
|
||||
exit 1
|
||||
else
|
||||
if [ -d "${arg}" ]; then
|
||||
echo "list-markers: file argument is a directory" >&2
|
||||
exit 1
|
||||
fi
|
||||
fi
|
||||
|
||||
echo -e "\n${arg}\n"
|
||||
mrks=$(cat ${arg} | grep -a "|[A-Z][a-zA-Z]," -o)
|
||||
mrksnum=$(echo "${mrks}" | wc -l)
|
||||
echo -e "(${mrksnum})\n${mrks}"
|
||||
}
|
||||
|
||||
#ls ${dir} | while read fn; do echo $fn; cat ${dir}$fn | xxd | head -n10 | tail -n3; done;
|
||||
if [ -f "${fildir}" ]; then
|
||||
|
||||
echo "analyzing single file ${fildir} ..."
|
||||
list-markers "${fildir}"
|
||||
|
||||
elif [ -d "${fildir}" ]; then
|
||||
|
||||
echo "analyzing all *.raw files in directory ${fildir} ..."
|
||||
lsfls=$(ls ${fildir}/*.raw | sed 's/\/\//\//g')
|
||||
echo -e "\n${lsfls}"
|
||||
|
||||
for fl in ${lsfls}; do
|
||||
list-markers "${fl}"
|
||||
done
|
||||
|
||||
else
|
||||
|
||||
echo "does not exist: ${fildir}" >&2
|
||||
exit 1
|
||||
|
||||
fi
|
||||
|
||||
ls ${dir} | while read fn; do echo $fn; cat ${dir}$fn | grep -a "|[A-Z][a-zA-Z]," -o | wc -l; done;
|
||||
ls ${dir} | while read fn; do echo $fn; cat ${dir}$fn | grep -a "|[A-Z][a-zA-Z]," -o; done;
|
||||
|
@@ -12,6 +12,7 @@
|
||||
|
||||
const std::string gittag("TAGSTRING");
|
||||
const std::string githash("HASHSTRING");
|
||||
const std::string timestamp("TIMESTAMPSTRING");
|
||||
|
||||
//---------------------------------------------------------------------------//
|
||||
|
||||
@@ -127,13 +128,13 @@ optkeys parse_args(int argc, char* argv[], bool list_args = false)
|
||||
|
||||
void show_version()
|
||||
{
|
||||
std::cout<<"imctermite ["<<gittag<<"-g"<<githash<<"]"<<"\n";
|
||||
std::cout<<"imctermite ["<<gittag<<"-g"<<githash<<"-"<<timestamp<<"]"<<"\n";
|
||||
}
|
||||
|
||||
void show_usage()
|
||||
{
|
||||
std::cout<<"\n"
|
||||
<<"imctermite ["<<gittag<<"-g"<<githash<<"] (github.com/RecordEvolution/IMCtermite.git)"
|
||||
<<"imctermite ["<<gittag<<"-g"<<githash<<"-"<<timestamp<<"] (https://github.com/RecordEvolution/IMCtermite.git)"
|
||||
<<"\n\n"
|
||||
<<"Decode IMC raw files and dump data as *.csv"
|
||||
<<"\n\n"
|
||||
|
Reference in New Issue
Block a user