Merge branch 'master' into develop
commit
db8e9eb556
|
@ -1,2 +1,2 @@
|
||||||
patreon: assimp
|
patreon: assimp
|
||||||
custom: https://www.paypal.com/cgi-bin/webscr?cmd=_s-xclick&hosted_button_id=4JRJVPXC4QJM4
|
open_collective: assimp
|
||||||
|
|
|
@ -18,6 +18,9 @@ build
|
||||||
*.VC.db-wal
|
*.VC.db-wal
|
||||||
*.VC.opendb
|
*.VC.opendb
|
||||||
*.ipch
|
*.ipch
|
||||||
|
.vs/
|
||||||
|
out/
|
||||||
|
CMakeSettings.json
|
||||||
|
|
||||||
# Output
|
# Output
|
||||||
bin/
|
bin/
|
||||||
|
|
|
@ -105,12 +105,6 @@ Become a financial contributor and help us sustain our community. [[Contribute](
|
||||||
Monthly donations via Patreon:
|
Monthly donations via Patreon:
|
||||||
<br>[![Patreon](https://cloud.githubusercontent.com/assets/8225057/5990484/70413560-a9ab-11e4-8942-1a63607c0b00.png)](http://www.patreon.com/assimp)
|
<br>[![Patreon](https://cloud.githubusercontent.com/assets/8225057/5990484/70413560-a9ab-11e4-8942-1a63607c0b00.png)](http://www.patreon.com/assimp)
|
||||||
|
|
||||||
<br>
|
|
||||||
|
|
||||||
One-off donations via PayPal:
|
|
||||||
<br>[![PayPal](https://www.paypalobjects.com/en_US/i/btn/btn_donate_LG.gif)](https://www.paypal.com/cgi-bin/webscr?cmd=_s-xclick&hosted_button_id=4JRJVPXC4QJM4)
|
|
||||||
|
|
||||||
<br>
|
|
||||||
|
|
||||||
#### Organizations
|
#### Organizations
|
||||||
|
|
||||||
|
|
81
appveyor.yml
81
appveyor.yml
|
@ -1,81 +0,0 @@
|
||||||
# AppVeyor file
|
|
||||||
# http://www.appveyor.com/docs/appveyor-yml
|
|
||||||
|
|
||||||
# clone directory
|
|
||||||
clone_folder: c:\projects\assimp
|
|
||||||
|
|
||||||
clone_depth: 1
|
|
||||||
|
|
||||||
# branches to build
|
|
||||||
branches:
|
|
||||||
# whitelist
|
|
||||||
only:
|
|
||||||
- master
|
|
||||||
|
|
||||||
matrix:
|
|
||||||
fast_finish: true
|
|
||||||
|
|
||||||
image:
|
|
||||||
- Visual Studio 2013
|
|
||||||
#- Visual Studio 2015
|
|
||||||
#- Visual Studio 2017
|
|
||||||
- Visual Studio 2019
|
|
||||||
#- MinGW
|
|
||||||
|
|
||||||
platform:
|
|
||||||
- Win32
|
|
||||||
- x64
|
|
||||||
|
|
||||||
configuration: Release
|
|
||||||
|
|
||||||
install:
|
|
||||||
- set PATH=C:\Ruby24-x64\bin;%PATH%
|
|
||||||
- set CMAKE_DEFINES -DASSIMP_WERROR=ON
|
|
||||||
- if [%COMPILER%]==[MinGW] set PATH=C:\MinGW\bin;%PATH%
|
|
||||||
- if "%APPVEYOR_BUILD_WORKER_IMAGE%"=="Visual Studio 2013" set CMAKE_GENERATOR_NAME=Visual Studio 12 2013
|
|
||||||
- if "%APPVEYOR_BUILD_WORKER_IMAGE%"=="Visual Studio 2015" set CMAKE_GENERATOR_NAME=Visual Studio 14 2015
|
|
||||||
- if "%APPVEYOR_BUILD_WORKER_IMAGE%"=="Visual Studio 2017" set CMAKE_GENERATOR_NAME=Visual Studio 15 2017
|
|
||||||
- if "%APPVEYOR_BUILD_WORKER_IMAGE%"=="Visual Studio 2019" set CMAKE_GENERATOR_NAME=Visual Studio 16 2019
|
|
||||||
- cmake %CMAKE_DEFINES% -G "%CMAKE_GENERATOR_NAME%" -A %platform% .
|
|
||||||
# Rename sh.exe as sh.exe in PATH interferes with MinGW - if "%APPVEYOR_BUILD_WORKER_IMAGE%"=="Visual Studio 2015" set CMAKE_GENERATOR_NAME=Visual Studio 14 2015
|
|
||||||
|
|
||||||
- rename "C:\Program Files\Git\usr\bin\sh.exe" "sh2.exe"
|
|
||||||
- set PATH=%PATH%;"C:\\Program Files (x86)\\Inno Setup 5"
|
|
||||||
- ps: Invoke-WebRequest -Uri https://download.microsoft.com/download/5/7/b/57b2947c-7221-4f33-b35e-2fc78cb10df4/vc_redist.x64.exe -OutFile .\packaging\windows-innosetup\vc_redist.x64.exe
|
|
||||||
- ps: Invoke-WebRequest -Uri https://download.microsoft.com/download/1/d/8/1d8137db-b5bb-4925-8c5d-927424a2e4de/vc_redist.x86.exe -OutFile .\packaging\windows-innosetup\vc_redist.x86.exe
|
|
||||||
|
|
||||||
cache:
|
|
||||||
- code\assimp.dir\%CONFIGURATION%
|
|
||||||
- contrib\zlib\zlibstatic.dir\%CONFIGURATION%
|
|
||||||
- contrib\zlib\zlib.dir\%CONFIGURATION%
|
|
||||||
- tools\assimp_cmd\assimp_cmd.dir\%CONFIGURATION%
|
|
||||||
- tools\assimp_view\assimp_viewer.dir\%CONFIGURATION%
|
|
||||||
- test\unit.dir\%CONFIGURATION%
|
|
||||||
- bin\.mtime_cache
|
|
||||||
|
|
||||||
before_build:
|
|
||||||
- echo NUMBER_OF_PROCESSORS=%NUMBER_OF_PROCESSORS%
|
|
||||||
- ruby scripts\AppVeyor\mtime_cache -g scripts\AppVeyor\cacheglobs.txt -c bin\.mtime_cache\cache.json
|
|
||||||
|
|
||||||
build_script:
|
|
||||||
cmake --build . --config Release -- /maxcpucount:2
|
|
||||||
|
|
||||||
after_build:
|
|
||||||
- if "%APPVEYOR_BUILD_WORKER_IMAGE%"=="Visual Studio 2017" (
|
|
||||||
if "%platform%"=="x64" (
|
|
||||||
iscc packaging\windows-innosetup\script_x64.iss
|
|
||||||
) else (
|
|
||||||
iscc packaging\windows-innosetup\script_x86.iss
|
|
||||||
)
|
|
||||||
)
|
|
||||||
- 7z a assimp.7z bin\%CONFIGURATION%\* lib\%CONFIGURATION%\*
|
|
||||||
|
|
||||||
test_script:
|
|
||||||
- cmd: bin\%CONFIGURATION%\unit.exe --gtest_output=xml:testout.xml
|
|
||||||
|
|
||||||
on_finish:
|
|
||||||
- ps: (new-object net.webclient).UploadFile("https://ci.appveyor.com/api/testresults/junit/$($env:APPVEYOR_JOB_ID)", (Resolve-Path .\testout.xml))
|
|
||||||
|
|
||||||
artifacts:
|
|
||||||
- path: assimp.7z
|
|
||||||
name: assimp_lib
|
|
|
@ -329,8 +329,8 @@ void AMFImporter::Postprocess_AddMetadata(const AMFMetaDataArray &metadataList,
|
||||||
sceneNode.mMetaData = aiMetadata::Alloc(static_cast<unsigned int>(metadataList.size()));
|
sceneNode.mMetaData = aiMetadata::Alloc(static_cast<unsigned int>(metadataList.size()));
|
||||||
size_t meta_idx(0);
|
size_t meta_idx(0);
|
||||||
|
|
||||||
for (const AMFMetadata &metadata : metadataList) {
|
for (const AMFMetadata *metadata : metadataList) {
|
||||||
sceneNode.mMetaData->Set(static_cast<unsigned int>(meta_idx++), metadata.Type, aiString(metadata.Value));
|
sceneNode.mMetaData->Set(static_cast<unsigned int>(meta_idx++), metadata->Type, aiString(metadata->Value));
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -334,7 +334,7 @@ void ColladaParser::ReadAssetInfo(XmlNode &node) {
|
||||||
const std::string ¤tName = currentNode.name();
|
const std::string ¤tName = currentNode.name();
|
||||||
if (currentName == "unit") {
|
if (currentName == "unit") {
|
||||||
mUnitSize = 1.f;
|
mUnitSize = 1.f;
|
||||||
XmlParser::getFloatAttribute(node, "meter", mUnitSize);
|
XmlParser::getFloatAttribute(currentNode, "meter", mUnitSize);
|
||||||
} else if (currentName == "up_axis") {
|
} else if (currentName == "up_axis") {
|
||||||
std::string v;
|
std::string v;
|
||||||
if (!XmlParser::getValueAsString(currentNode, v)) {
|
if (!XmlParser::getValueAsString(currentNode, v)) {
|
||||||
|
@ -459,7 +459,6 @@ void ColladaParser::PostProcessRootAnimations() {
|
||||||
|
|
||||||
if (animation != mAnimationLibrary.end()) {
|
if (animation != mAnimationLibrary.end()) {
|
||||||
Animation *pSourceAnimation = animation->second;
|
Animation *pSourceAnimation = animation->second;
|
||||||
|
|
||||||
pSourceAnimation->CollectChannelsRecursively(clip->mChannels);
|
pSourceAnimation->CollectChannelsRecursively(clip->mChannels);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -1738,14 +1737,16 @@ size_t ColladaParser::ReadPrimitives(XmlNode &node, Mesh &pMesh, std::vector<Inp
|
||||||
|
|
||||||
// and read all indices into a temporary array
|
// and read all indices into a temporary array
|
||||||
std::vector<size_t> indices;
|
std::vector<size_t> indices;
|
||||||
if (expectedPointCount > 0)
|
if (expectedPointCount > 0) {
|
||||||
indices.reserve(expectedPointCount * numOffsets);
|
indices.reserve(expectedPointCount * numOffsets);
|
||||||
|
}
|
||||||
|
|
||||||
if (pNumPrimitives > 0) // It is possible to not contain any indices
|
// It is possible to not contain any indices
|
||||||
{
|
if (pNumPrimitives > 0) {
|
||||||
std::string v;
|
std::string v;
|
||||||
XmlParser::getValueAsString(node, v);
|
XmlParser::getValueAsString(node, v);
|
||||||
const char *content = v.c_str();
|
const char *content = v.c_str();
|
||||||
|
SkipSpacesAndLineEnd(&content);
|
||||||
while (*content != 0) {
|
while (*content != 0) {
|
||||||
// read a value.
|
// read a value.
|
||||||
// Hack: (thom) Some exporters put negative indices sometimes. We just try to carry on anyways.
|
// Hack: (thom) Some exporters put negative indices sometimes. We just try to carry on anyways.
|
||||||
|
@ -1772,21 +1773,24 @@ size_t ColladaParser::ReadPrimitives(XmlNode &node, Mesh &pMesh, std::vector<Inp
|
||||||
// find the data for all sources
|
// find the data for all sources
|
||||||
for (std::vector<InputChannel>::iterator it = pMesh.mPerVertexData.begin(); it != pMesh.mPerVertexData.end(); ++it) {
|
for (std::vector<InputChannel>::iterator it = pMesh.mPerVertexData.begin(); it != pMesh.mPerVertexData.end(); ++it) {
|
||||||
InputChannel &input = *it;
|
InputChannel &input = *it;
|
||||||
if (input.mResolved)
|
if (input.mResolved) {
|
||||||
continue;
|
continue;
|
||||||
|
}
|
||||||
|
|
||||||
// find accessor
|
// find accessor
|
||||||
input.mResolved = &ResolveLibraryReference(mAccessorLibrary, input.mAccessor);
|
input.mResolved = &ResolveLibraryReference(mAccessorLibrary, input.mAccessor);
|
||||||
// resolve accessor's data pointer as well, if necessary
|
// resolve accessor's data pointer as well, if necessary
|
||||||
const Accessor *acc = input.mResolved;
|
const Accessor *acc = input.mResolved;
|
||||||
if (!acc->mData)
|
if (!acc->mData) {
|
||||||
acc->mData = &ResolveLibraryReference(mDataLibrary, acc->mSource);
|
acc->mData = &ResolveLibraryReference(mDataLibrary, acc->mSource);
|
||||||
}
|
}
|
||||||
|
}
|
||||||
// and the same for the per-index channels
|
// and the same for the per-index channels
|
||||||
for (std::vector<InputChannel>::iterator it = pPerIndexChannels.begin(); it != pPerIndexChannels.end(); ++it) {
|
for (std::vector<InputChannel>::iterator it = pPerIndexChannels.begin(); it != pPerIndexChannels.end(); ++it) {
|
||||||
InputChannel &input = *it;
|
InputChannel &input = *it;
|
||||||
if (input.mResolved)
|
if (input.mResolved) {
|
||||||
continue;
|
continue;
|
||||||
|
}
|
||||||
|
|
||||||
// ignore vertex pointer, it doesn't refer to an accessor
|
// ignore vertex pointer, it doesn't refer to an accessor
|
||||||
if (input.mType == IT_Vertex) {
|
if (input.mType == IT_Vertex) {
|
||||||
|
@ -1801,9 +1805,10 @@ size_t ColladaParser::ReadPrimitives(XmlNode &node, Mesh &pMesh, std::vector<Inp
|
||||||
input.mResolved = &ResolveLibraryReference(mAccessorLibrary, input.mAccessor);
|
input.mResolved = &ResolveLibraryReference(mAccessorLibrary, input.mAccessor);
|
||||||
// resolve accessor's data pointer as well, if necessary
|
// resolve accessor's data pointer as well, if necessary
|
||||||
const Accessor *acc = input.mResolved;
|
const Accessor *acc = input.mResolved;
|
||||||
if (!acc->mData)
|
if (!acc->mData) {
|
||||||
acc->mData = &ResolveLibraryReference(mDataLibrary, acc->mSource);
|
acc->mData = &ResolveLibraryReference(mDataLibrary, acc->mSource);
|
||||||
}
|
}
|
||||||
|
}
|
||||||
|
|
||||||
// For continued primitives, the given count does not come all in one <p>, but only one primitive per <p>
|
// For continued primitives, the given count does not come all in one <p>, but only one primitive per <p>
|
||||||
size_t numPrimitives = pNumPrimitives;
|
size_t numPrimitives = pNumPrimitives;
|
||||||
|
@ -1884,11 +1889,13 @@ void ColladaParser::CopyVertex(size_t currentVertex, size_t numOffsets, size_t n
|
||||||
ai_assert((baseOffset + numOffsets - 1) < indices.size());
|
ai_assert((baseOffset + numOffsets - 1) < indices.size());
|
||||||
|
|
||||||
// extract per-vertex channels using the global per-vertex offset
|
// extract per-vertex channels using the global per-vertex offset
|
||||||
for (std::vector<InputChannel>::iterator it = pMesh.mPerVertexData.begin(); it != pMesh.mPerVertexData.end(); ++it)
|
for (std::vector<InputChannel>::iterator it = pMesh.mPerVertexData.begin(); it != pMesh.mPerVertexData.end(); ++it) {
|
||||||
ExtractDataObjectFromChannel(*it, indices[baseOffset + perVertexOffset], pMesh);
|
ExtractDataObjectFromChannel(*it, indices[baseOffset + perVertexOffset], pMesh);
|
||||||
|
}
|
||||||
// and extract per-index channels using there specified offset
|
// and extract per-index channels using there specified offset
|
||||||
for (std::vector<InputChannel>::iterator it = pPerIndexChannels.begin(); it != pPerIndexChannels.end(); ++it)
|
for (std::vector<InputChannel>::iterator it = pPerIndexChannels.begin(); it != pPerIndexChannels.end(); ++it) {
|
||||||
ExtractDataObjectFromChannel(*it, indices[baseOffset + it->mOffset], pMesh);
|
ExtractDataObjectFromChannel(*it, indices[baseOffset + it->mOffset], pMesh);
|
||||||
|
}
|
||||||
|
|
||||||
// store the vertex-data index for later assignment of bone vertex weights
|
// store the vertex-data index for later assignment of bone vertex weights
|
||||||
pMesh.mFacePosIndices.push_back(indices[baseOffset + perVertexOffset]);
|
pMesh.mFacePosIndices.push_back(indices[baseOffset + perVertexOffset]);
|
||||||
|
@ -1912,8 +1919,9 @@ void ColladaParser::ReadPrimTriStrips(size_t numOffsets, size_t perVertexOffset,
|
||||||
// Extracts a single object from an input channel and stores it in the appropriate mesh data array
|
// Extracts a single object from an input channel and stores it in the appropriate mesh data array
|
||||||
void ColladaParser::ExtractDataObjectFromChannel(const InputChannel &pInput, size_t pLocalIndex, Mesh &pMesh) {
|
void ColladaParser::ExtractDataObjectFromChannel(const InputChannel &pInput, size_t pLocalIndex, Mesh &pMesh) {
|
||||||
// ignore vertex referrer - we handle them that separate
|
// ignore vertex referrer - we handle them that separate
|
||||||
if (pInput.mType == IT_Vertex)
|
if (pInput.mType == IT_Vertex) {
|
||||||
return;
|
return;
|
||||||
|
}
|
||||||
|
|
||||||
const Accessor &acc = *pInput.mResolved;
|
const Accessor &acc = *pInput.mResolved;
|
||||||
if (pLocalIndex >= acc.mCount) {
|
if (pLocalIndex >= acc.mCount) {
|
||||||
|
@ -1926,16 +1934,18 @@ void ColladaParser::ExtractDataObjectFromChannel(const InputChannel &pInput, siz
|
||||||
// assemble according to the accessors component sub-offset list. We don't care, yet,
|
// assemble according to the accessors component sub-offset list. We don't care, yet,
|
||||||
// what kind of object exactly we're extracting here
|
// what kind of object exactly we're extracting here
|
||||||
ai_real obj[4];
|
ai_real obj[4];
|
||||||
for (size_t c = 0; c < 4; ++c)
|
for (size_t c = 0; c < 4; ++c) {
|
||||||
obj[c] = dataObject[acc.mSubOffset[c]];
|
obj[c] = dataObject[acc.mSubOffset[c]];
|
||||||
|
}
|
||||||
|
|
||||||
// now we reinterpret it according to the type we're reading here
|
// now we reinterpret it according to the type we're reading here
|
||||||
switch (pInput.mType) {
|
switch (pInput.mType) {
|
||||||
case IT_Position: // ignore all position streams except 0 - there can be only one position
|
case IT_Position: // ignore all position streams except 0 - there can be only one position
|
||||||
if (pInput.mIndex == 0)
|
if (pInput.mIndex == 0) {
|
||||||
pMesh.mPositions.push_back(aiVector3D(obj[0], obj[1], obj[2]));
|
pMesh.mPositions.push_back(aiVector3D(obj[0], obj[1], obj[2]));
|
||||||
else
|
} else {
|
||||||
ASSIMP_LOG_ERROR("Collada: just one vertex position stream supported");
|
ASSIMP_LOG_ERROR("Collada: just one vertex position stream supported");
|
||||||
|
}
|
||||||
break;
|
break;
|
||||||
case IT_Normal:
|
case IT_Normal:
|
||||||
// pad to current vertex count if necessary
|
// pad to current vertex count if necessary
|
||||||
|
@ -1943,10 +1953,11 @@ void ColladaParser::ExtractDataObjectFromChannel(const InputChannel &pInput, siz
|
||||||
pMesh.mNormals.insert(pMesh.mNormals.end(), pMesh.mPositions.size() - pMesh.mNormals.size() - 1, aiVector3D(0, 1, 0));
|
pMesh.mNormals.insert(pMesh.mNormals.end(), pMesh.mPositions.size() - pMesh.mNormals.size() - 1, aiVector3D(0, 1, 0));
|
||||||
|
|
||||||
// ignore all normal streams except 0 - there can be only one normal
|
// ignore all normal streams except 0 - there can be only one normal
|
||||||
if (pInput.mIndex == 0)
|
if (pInput.mIndex == 0) {
|
||||||
pMesh.mNormals.push_back(aiVector3D(obj[0], obj[1], obj[2]));
|
pMesh.mNormals.push_back(aiVector3D(obj[0], obj[1], obj[2]));
|
||||||
else
|
} else {
|
||||||
ASSIMP_LOG_ERROR("Collada: just one vertex normal stream supported");
|
ASSIMP_LOG_ERROR("Collada: just one vertex normal stream supported");
|
||||||
|
}
|
||||||
break;
|
break;
|
||||||
case IT_Tangent:
|
case IT_Tangent:
|
||||||
// pad to current vertex count if necessary
|
// pad to current vertex count if necessary
|
||||||
|
@ -1954,21 +1965,24 @@ void ColladaParser::ExtractDataObjectFromChannel(const InputChannel &pInput, siz
|
||||||
pMesh.mTangents.insert(pMesh.mTangents.end(), pMesh.mPositions.size() - pMesh.mTangents.size() - 1, aiVector3D(1, 0, 0));
|
pMesh.mTangents.insert(pMesh.mTangents.end(), pMesh.mPositions.size() - pMesh.mTangents.size() - 1, aiVector3D(1, 0, 0));
|
||||||
|
|
||||||
// ignore all tangent streams except 0 - there can be only one tangent
|
// ignore all tangent streams except 0 - there can be only one tangent
|
||||||
if (pInput.mIndex == 0)
|
if (pInput.mIndex == 0) {
|
||||||
pMesh.mTangents.push_back(aiVector3D(obj[0], obj[1], obj[2]));
|
pMesh.mTangents.push_back(aiVector3D(obj[0], obj[1], obj[2]));
|
||||||
else
|
} else {
|
||||||
ASSIMP_LOG_ERROR("Collada: just one vertex tangent stream supported");
|
ASSIMP_LOG_ERROR("Collada: just one vertex tangent stream supported");
|
||||||
|
}
|
||||||
break;
|
break;
|
||||||
case IT_Bitangent:
|
case IT_Bitangent:
|
||||||
// pad to current vertex count if necessary
|
// pad to current vertex count if necessary
|
||||||
if (pMesh.mBitangents.size() < pMesh.mPositions.size() - 1)
|
if (pMesh.mBitangents.size() < pMesh.mPositions.size() - 1) {
|
||||||
pMesh.mBitangents.insert(pMesh.mBitangents.end(), pMesh.mPositions.size() - pMesh.mBitangents.size() - 1, aiVector3D(0, 0, 1));
|
pMesh.mBitangents.insert(pMesh.mBitangents.end(), pMesh.mPositions.size() - pMesh.mBitangents.size() - 1, aiVector3D(0, 0, 1));
|
||||||
|
}
|
||||||
|
|
||||||
// ignore all bitangent streams except 0 - there can be only one bitangent
|
// ignore all bitangent streams except 0 - there can be only one bitangent
|
||||||
if (pInput.mIndex == 0)
|
if (pInput.mIndex == 0) {
|
||||||
pMesh.mBitangents.push_back(aiVector3D(obj[0], obj[1], obj[2]));
|
pMesh.mBitangents.push_back(aiVector3D(obj[0], obj[1], obj[2]));
|
||||||
else
|
} else {
|
||||||
ASSIMP_LOG_ERROR("Collada: just one vertex bitangent stream supported");
|
ASSIMP_LOG_ERROR("Collada: just one vertex bitangent stream supported");
|
||||||
|
}
|
||||||
break;
|
break;
|
||||||
case IT_Texcoord:
|
case IT_Texcoord:
|
||||||
// up to 4 texture coord sets are fine, ignore the others
|
// up to 4 texture coord sets are fine, ignore the others
|
||||||
|
@ -1979,8 +1993,9 @@ void ColladaParser::ExtractDataObjectFromChannel(const InputChannel &pInput, siz
|
||||||
pMesh.mPositions.size() - pMesh.mTexCoords[pInput.mIndex].size() - 1, aiVector3D(0, 0, 0));
|
pMesh.mPositions.size() - pMesh.mTexCoords[pInput.mIndex].size() - 1, aiVector3D(0, 0, 0));
|
||||||
|
|
||||||
pMesh.mTexCoords[pInput.mIndex].push_back(aiVector3D(obj[0], obj[1], obj[2]));
|
pMesh.mTexCoords[pInput.mIndex].push_back(aiVector3D(obj[0], obj[1], obj[2]));
|
||||||
if (0 != acc.mSubOffset[2] || 0 != acc.mSubOffset[3]) /* hack ... consider cleaner solution */
|
if (0 != acc.mSubOffset[2] || 0 != acc.mSubOffset[3]) {
|
||||||
pMesh.mNumUVComponents[pInput.mIndex] = 3;
|
pMesh.mNumUVComponents[pInput.mIndex] = 3;
|
||||||
|
}
|
||||||
} else {
|
} else {
|
||||||
ASSIMP_LOG_ERROR("Collada: too many texture coordinate sets. Skipping.");
|
ASSIMP_LOG_ERROR("Collada: too many texture coordinate sets. Skipping.");
|
||||||
}
|
}
|
||||||
|
|
|
@ -54,7 +54,6 @@ OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
||||||
#include <assimp/ByteSwapper.h>
|
#include <assimp/ByteSwapper.h>
|
||||||
#include <assimp/ParsingUtils.h>
|
#include <assimp/ParsingUtils.h>
|
||||||
|
|
||||||
#include <algorithm> // std::transform
|
|
||||||
#include "FBXUtil.h"
|
#include "FBXUtil.h"
|
||||||
|
|
||||||
namespace Assimp {
|
namespace Assimp {
|
||||||
|
@ -63,9 +62,8 @@ namespace FBX {
|
||||||
using namespace Util;
|
using namespace Util;
|
||||||
|
|
||||||
// ------------------------------------------------------------------------------------------------
|
// ------------------------------------------------------------------------------------------------
|
||||||
Material::Material(uint64_t id, const Element& element, const Document& doc, const std::string& name)
|
Material::Material(uint64_t id, const Element& element, const Document& doc, const std::string& name) :
|
||||||
: Object(id,element,name)
|
Object(id,element,name) {
|
||||||
{
|
|
||||||
const Scope& sc = GetRequiredScope(element);
|
const Scope& sc = GetRequiredScope(element);
|
||||||
|
|
||||||
const Element* const ShadingModel = sc["ShadingModel"];
|
const Element* const ShadingModel = sc["ShadingModel"];
|
||||||
|
@ -77,23 +75,21 @@ Material::Material(uint64_t id, const Element& element, const Document& doc, con
|
||||||
|
|
||||||
if(ShadingModel) {
|
if(ShadingModel) {
|
||||||
shading = ParseTokenAsString(GetRequiredToken(*ShadingModel,0));
|
shading = ParseTokenAsString(GetRequiredToken(*ShadingModel,0));
|
||||||
}
|
} else {
|
||||||
else {
|
|
||||||
DOMWarning("shading mode not specified, assuming phong",&element);
|
DOMWarning("shading mode not specified, assuming phong",&element);
|
||||||
shading = "phong";
|
shading = "phong";
|
||||||
}
|
}
|
||||||
|
|
||||||
std::string templateName;
|
|
||||||
|
|
||||||
// lower-case shading because Blender (for example) writes "Phong"
|
// lower-case shading because Blender (for example) writes "Phong"
|
||||||
std::transform(shading.data(), shading.data() + shading.size(), std::addressof(shading[0]), Assimp::ToLower<char>);
|
for (size_t i = 0; i < shading.length(); ++i) {
|
||||||
|
shading[i] = static_cast<char>(tolower(shading[i]));
|
||||||
|
}
|
||||||
|
std::string templateName;
|
||||||
if(shading == "phong") {
|
if(shading == "phong") {
|
||||||
templateName = "Material.FbxSurfacePhong";
|
templateName = "Material.FbxSurfacePhong";
|
||||||
}
|
} else if(shading == "lambert") {
|
||||||
else if(shading == "lambert") {
|
|
||||||
templateName = "Material.FbxSurfaceLambert";
|
templateName = "Material.FbxSurfaceLambert";
|
||||||
}
|
} else {
|
||||||
else {
|
|
||||||
DOMWarning("shading mode not recognized: " + shading,&element);
|
DOMWarning("shading mode not recognized: " + shading,&element);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -102,20 +98,19 @@ Material::Material(uint64_t id, const Element& element, const Document& doc, con
|
||||||
// resolve texture links
|
// resolve texture links
|
||||||
const std::vector<const Connection*>& conns = doc.GetConnectionsByDestinationSequenced(ID());
|
const std::vector<const Connection*>& conns = doc.GetConnectionsByDestinationSequenced(ID());
|
||||||
for(const Connection* con : conns) {
|
for(const Connection* con : conns) {
|
||||||
|
|
||||||
// texture link to properties, not objects
|
// texture link to properties, not objects
|
||||||
if (!con->PropertyName().length()) {
|
if ( 0 == con->PropertyName().length()) {
|
||||||
continue;
|
continue;
|
||||||
}
|
}
|
||||||
|
|
||||||
const Object* const ob = con->SourceObject();
|
const Object* const ob = con->SourceObject();
|
||||||
if(!ob) {
|
if(nullptr == ob) {
|
||||||
DOMWarning("failed to read source object for texture link, ignoring",&element);
|
DOMWarning("failed to read source object for texture link, ignoring",&element);
|
||||||
continue;
|
continue;
|
||||||
}
|
}
|
||||||
|
|
||||||
const Texture* const tex = dynamic_cast<const Texture*>(ob);
|
const Texture* const tex = dynamic_cast<const Texture*>(ob);
|
||||||
if(!tex) {
|
if(nullptr == tex) {
|
||||||
const LayeredTexture* const layeredTexture = dynamic_cast<const LayeredTexture*>(ob);
|
const LayeredTexture* const layeredTexture = dynamic_cast<const LayeredTexture*>(ob);
|
||||||
if(!layeredTexture) {
|
if(!layeredTexture) {
|
||||||
DOMWarning("source object for texture link is not a texture or layered texture, ignoring",&element);
|
DOMWarning("source object for texture link is not a texture or layered texture, ignoring",&element);
|
||||||
|
@ -128,9 +123,7 @@ Material::Material(uint64_t id, const Element& element, const Document& doc, con
|
||||||
|
|
||||||
layeredTextures[prop] = layeredTexture;
|
layeredTextures[prop] = layeredTexture;
|
||||||
((LayeredTexture*)layeredTexture)->fillTexture(doc);
|
((LayeredTexture*)layeredTexture)->fillTexture(doc);
|
||||||
}
|
} else {
|
||||||
else
|
|
||||||
{
|
|
||||||
const std::string& prop = con->PropertyName();
|
const std::string& prop = con->PropertyName();
|
||||||
if (textures.find(prop) != textures.end()) {
|
if (textures.find(prop) != textures.end()) {
|
||||||
DOMWarning("duplicate texture link: " + prop,&element);
|
DOMWarning("duplicate texture link: " + prop,&element);
|
||||||
|
@ -138,23 +131,20 @@ Material::Material(uint64_t id, const Element& element, const Document& doc, con
|
||||||
|
|
||||||
textures[prop] = tex;
|
textures[prop] = tex;
|
||||||
}
|
}
|
||||||
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
// ------------------------------------------------------------------------------------------------
|
// ------------------------------------------------------------------------------------------------
|
||||||
Material::~Material()
|
Material::~Material() {
|
||||||
{
|
// empty
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
// ------------------------------------------------------------------------------------------------
|
// ------------------------------------------------------------------------------------------------
|
||||||
Texture::Texture(uint64_t id, const Element& element, const Document& doc, const std::string& name)
|
Texture::Texture(uint64_t id, const Element& element, const Document& doc, const std::string& name) :
|
||||||
: Object(id,element,name)
|
Object(id,element,name),
|
||||||
, uvScaling(1.0f,1.0f)
|
uvScaling(1.0f,1.0f),
|
||||||
, media(0)
|
media(0) {
|
||||||
{
|
|
||||||
const Scope& sc = GetRequiredScope(element);
|
const Scope& sc = GetRequiredScope(element);
|
||||||
|
|
||||||
const Element* const Type = sc["Type"];
|
const Element* const Type = sc["Type"];
|
||||||
|
@ -194,8 +184,7 @@ Texture::Texture(uint64_t id, const Element& element, const Document& doc, const
|
||||||
crop[1] = ParseTokenAsInt(GetRequiredToken(*Cropping,1));
|
crop[1] = ParseTokenAsInt(GetRequiredToken(*Cropping,1));
|
||||||
crop[2] = ParseTokenAsInt(GetRequiredToken(*Cropping,2));
|
crop[2] = ParseTokenAsInt(GetRequiredToken(*Cropping,2));
|
||||||
crop[3] = ParseTokenAsInt(GetRequiredToken(*Cropping,3));
|
crop[3] = ParseTokenAsInt(GetRequiredToken(*Cropping,3));
|
||||||
}
|
} else {
|
||||||
else {
|
|
||||||
// vc8 doesn't support the crop() syntax in initialization lists
|
// vc8 doesn't support the crop() syntax in initialization lists
|
||||||
// (and vc9 WARNS about the new (i.e. compliant) behaviour).
|
// (and vc9 WARNS about the new (i.e. compliant) behaviour).
|
||||||
crop[0] = crop[1] = crop[2] = crop[3] = 0;
|
crop[0] = crop[1] = crop[2] = crop[3] = 0;
|
||||||
|
@ -226,7 +215,7 @@ Texture::Texture(uint64_t id, const Element& element, const Document& doc, const
|
||||||
const std::vector<const Connection*>& conns = doc.GetConnectionsByDestinationSequenced(ID());
|
const std::vector<const Connection*>& conns = doc.GetConnectionsByDestinationSequenced(ID());
|
||||||
for(const Connection* con : conns) {
|
for(const Connection* con : conns) {
|
||||||
const Object* const ob = con->SourceObject();
|
const Object* const ob = con->SourceObject();
|
||||||
if(!ob) {
|
if (nullptr == ob) {
|
||||||
DOMWarning("failed to read source object for texture link, ignoring",&element);
|
DOMWarning("failed to read source object for texture link, ignoring",&element);
|
||||||
continue;
|
continue;
|
||||||
}
|
}
|
||||||
|
@ -240,46 +229,38 @@ Texture::Texture(uint64_t id, const Element& element, const Document& doc, const
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
Texture::~Texture()
|
Texture::~Texture() {
|
||||||
{
|
// empty
|
||||||
|
|
||||||
}
|
}
|
||||||
|
|
||||||
LayeredTexture::LayeredTexture(uint64_t id, const Element& element, const Document& /*doc*/, const std::string& name)
|
LayeredTexture::LayeredTexture(uint64_t id, const Element& element, const Document& /*doc*/, const std::string& name) :
|
||||||
: Object(id,element,name)
|
Object(id,element,name),
|
||||||
,blendMode(BlendMode_Modulate)
|
blendMode(BlendMode_Modulate),
|
||||||
,alpha(1)
|
alpha(1) {
|
||||||
{
|
|
||||||
const Scope& sc = GetRequiredScope(element);
|
const Scope& sc = GetRequiredScope(element);
|
||||||
|
|
||||||
const Element* const BlendModes = sc["BlendModes"];
|
const Element* const BlendModes = sc["BlendModes"];
|
||||||
const Element* const Alphas = sc["Alphas"];
|
const Element* const Alphas = sc["Alphas"];
|
||||||
|
|
||||||
|
if (nullptr != BlendModes) {
|
||||||
if(BlendModes!=0)
|
|
||||||
{
|
|
||||||
blendMode = (BlendMode)ParseTokenAsInt(GetRequiredToken(*BlendModes,0));
|
blendMode = (BlendMode)ParseTokenAsInt(GetRequiredToken(*BlendModes,0));
|
||||||
}
|
}
|
||||||
if(Alphas!=0)
|
if (nullptr != Alphas) {
|
||||||
{
|
|
||||||
alpha = ParseTokenAsFloat(GetRequiredToken(*Alphas,0));
|
alpha = ParseTokenAsFloat(GetRequiredToken(*Alphas,0));
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
LayeredTexture::~LayeredTexture()
|
LayeredTexture::~LayeredTexture() {
|
||||||
{
|
// empty
|
||||||
|
|
||||||
}
|
}
|
||||||
|
|
||||||
void LayeredTexture::fillTexture(const Document& doc)
|
void LayeredTexture::fillTexture(const Document& doc) {
|
||||||
{
|
|
||||||
const std::vector<const Connection*>& conns = doc.GetConnectionsByDestinationSequenced(ID());
|
const std::vector<const Connection*>& conns = doc.GetConnectionsByDestinationSequenced(ID());
|
||||||
for(size_t i = 0; i < conns.size();++i)
|
for(size_t i = 0; i < conns.size();++i) {
|
||||||
{
|
|
||||||
const Connection* con = conns.at(i);
|
const Connection* con = conns.at(i);
|
||||||
|
|
||||||
const Object* const ob = con->SourceObject();
|
const Object* const ob = con->SourceObject();
|
||||||
if(!ob) {
|
if (nullptr == ob) {
|
||||||
DOMWarning("failed to read source object for texture link, ignoring",&element);
|
DOMWarning("failed to read source object for texture link, ignoring",&element);
|
||||||
continue;
|
continue;
|
||||||
}
|
}
|
||||||
|
@ -290,13 +271,11 @@ void LayeredTexture::fillTexture(const Document& doc)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
// ------------------------------------------------------------------------------------------------
|
// ------------------------------------------------------------------------------------------------
|
||||||
Video::Video(uint64_t id, const Element& element, const Document& doc, const std::string& name)
|
Video::Video(uint64_t id, const Element& element, const Document& doc, const std::string& name) :
|
||||||
: Object(id,element,name)
|
Object(id,element,name),
|
||||||
, contentLength(0)
|
contentLength(0),
|
||||||
, content(0)
|
content(0) {
|
||||||
{
|
|
||||||
const Scope& sc = GetRequiredScope(element);
|
const Scope& sc = GetRequiredScope(element);
|
||||||
|
|
||||||
const Element* const Type = sc["Type"];
|
const Element* const Type = sc["Type"];
|
||||||
|
@ -324,52 +303,43 @@ Video::Video(uint64_t id, const Element& element, const Document& doc, const std
|
||||||
if (!token.IsBinary()) {
|
if (!token.IsBinary()) {
|
||||||
if (*data != '"') {
|
if (*data != '"') {
|
||||||
DOMError("embedded content is not surrounded by quotation marks", &element);
|
DOMError("embedded content is not surrounded by quotation marks", &element);
|
||||||
}
|
} else {
|
||||||
else {
|
|
||||||
size_t targetLength = 0;
|
size_t targetLength = 0;
|
||||||
auto numTokens = Content->Tokens().size();
|
auto numTokens = Content->Tokens().size();
|
||||||
// First time compute size (it could be large like 64Gb and it is good to allocate it once)
|
// First time compute size (it could be large like 64Gb and it is good to allocate it once)
|
||||||
for (uint32_t tokenIdx = 0; tokenIdx < numTokens; ++tokenIdx)
|
for (uint32_t tokenIdx = 0; tokenIdx < numTokens; ++tokenIdx) {
|
||||||
{
|
|
||||||
const Token& dataToken = GetRequiredToken(*Content, tokenIdx);
|
const Token& dataToken = GetRequiredToken(*Content, tokenIdx);
|
||||||
size_t tokenLength = dataToken.end() - dataToken.begin() - 2; // ignore double quotes
|
size_t tokenLength = dataToken.end() - dataToken.begin() - 2; // ignore double quotes
|
||||||
const char* base64data = dataToken.begin() + 1;
|
const char* base64data = dataToken.begin() + 1;
|
||||||
const size_t outLength = Util::ComputeDecodedSizeBase64(base64data, tokenLength);
|
const size_t outLength = Util::ComputeDecodedSizeBase64(base64data, tokenLength);
|
||||||
if (outLength == 0)
|
if (outLength == 0) {
|
||||||
{
|
|
||||||
DOMError("Corrupted embedded content found", &element);
|
DOMError("Corrupted embedded content found", &element);
|
||||||
}
|
}
|
||||||
targetLength += outLength;
|
targetLength += outLength;
|
||||||
}
|
}
|
||||||
if (targetLength == 0)
|
if (targetLength == 0) {
|
||||||
{
|
|
||||||
DOMError("Corrupted embedded content found", &element);
|
DOMError("Corrupted embedded content found", &element);
|
||||||
}
|
}
|
||||||
content = new uint8_t[targetLength];
|
content = new uint8_t[targetLength];
|
||||||
contentLength = static_cast<uint64_t>(targetLength);
|
contentLength = static_cast<uint64_t>(targetLength);
|
||||||
size_t dst_offset = 0;
|
size_t dst_offset = 0;
|
||||||
for (uint32_t tokenIdx = 0; tokenIdx < numTokens; ++tokenIdx)
|
for (uint32_t tokenIdx = 0; tokenIdx < numTokens; ++tokenIdx) {
|
||||||
{
|
|
||||||
const Token& dataToken = GetRequiredToken(*Content, tokenIdx);
|
const Token& dataToken = GetRequiredToken(*Content, tokenIdx);
|
||||||
size_t tokenLength = dataToken.end() - dataToken.begin() - 2; // ignore double quotes
|
size_t tokenLength = dataToken.end() - dataToken.begin() - 2; // ignore double quotes
|
||||||
const char* base64data = dataToken.begin() + 1;
|
const char* base64data = dataToken.begin() + 1;
|
||||||
dst_offset += Util::DecodeBase64(base64data, tokenLength, content + dst_offset, targetLength - dst_offset);
|
dst_offset += Util::DecodeBase64(base64data, tokenLength, content + dst_offset, targetLength - dst_offset);
|
||||||
}
|
}
|
||||||
if (targetLength != dst_offset)
|
if (targetLength != dst_offset) {
|
||||||
{
|
|
||||||
delete[] content;
|
delete[] content;
|
||||||
contentLength = 0;
|
contentLength = 0;
|
||||||
DOMError("Corrupted embedded content found", &element);
|
DOMError("Corrupted embedded content found", &element);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
} else if (static_cast<size_t>(token.end() - data) < 5) {
|
||||||
else if (static_cast<size_t>(token.end() - data) < 5) {
|
|
||||||
DOMError("binary data array is too short, need five (5) bytes for type signature and element count", &element);
|
DOMError("binary data array is too short, need five (5) bytes for type signature and element count", &element);
|
||||||
}
|
} else if (*data != 'R') {
|
||||||
else if (*data != 'R') {
|
|
||||||
DOMWarning("video content is not raw binary data, ignoring", &element);
|
DOMWarning("video content is not raw binary data, ignoring", &element);
|
||||||
}
|
} else {
|
||||||
else {
|
|
||||||
// read number of elements
|
// read number of elements
|
||||||
uint32_t len = 0;
|
uint32_t len = 0;
|
||||||
::memcpy(&len, data + 1, sizeof(len));
|
::memcpy(&len, data + 1, sizeof(len));
|
||||||
|
@ -380,8 +350,7 @@ Video::Video(uint64_t id, const Element& element, const Document& doc, const std
|
||||||
content = new uint8_t[len];
|
content = new uint8_t[len];
|
||||||
::memcpy(content, data + 5, len);
|
::memcpy(content, data + 5, len);
|
||||||
}
|
}
|
||||||
} catch (const runtime_error& runtimeError)
|
} catch (const runtime_error& runtimeError) {
|
||||||
{
|
|
||||||
//we don't need the content data for contents that has already been loaded
|
//we don't need the content data for contents that has already been loaded
|
||||||
ASSIMP_LOG_VERBOSE_DEBUG_F("Caught exception in FBXMaterial (likely because content was already loaded): ",
|
ASSIMP_LOG_VERBOSE_DEBUG_F("Caught exception in FBXMaterial (likely because content was already loaded): ",
|
||||||
runtimeError.what());
|
runtimeError.what());
|
||||||
|
@ -392,14 +361,11 @@ Video::Video(uint64_t id, const Element& element, const Document& doc, const std
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
Video::~Video()
|
Video::~Video() {
|
||||||
{
|
|
||||||
if(content) {
|
|
||||||
delete[] content;
|
delete[] content;
|
||||||
}
|
}
|
||||||
}
|
|
||||||
|
|
||||||
} //!FBX
|
} //!FBX
|
||||||
} //!Assimp
|
} //!Assimp
|
||||||
|
|
||||||
#endif
|
#endif // ASSIMP_BUILD_NO_FBX_IMPORTER
|
||||||
|
|
|
@ -1229,7 +1229,7 @@ unsigned int glTF2Exporter::ExportNode(const aiNode* n, Ref<Node>& parent)
|
||||||
node->name = name;
|
node->name = name;
|
||||||
|
|
||||||
if (!n->mTransformation.IsIdentity()) {
|
if (!n->mTransformation.IsIdentity()) {
|
||||||
if (mScene->mNumAnimations > 0) {
|
if (mScene->mNumAnimations > 0 || (mProperties && mProperties->HasPropertyBool("GLTF2_NODE_IN_TRS"))) {
|
||||||
aiQuaternion quaternion;
|
aiQuaternion quaternion;
|
||||||
n->mTransformation.Decompose(*reinterpret_cast<aiVector3D *>(&node->scale.value), quaternion, *reinterpret_cast<aiVector3D *>(&node->translation.value));
|
n->mTransformation.Decompose(*reinterpret_cast<aiVector3D *>(&node->scale.value), quaternion, *reinterpret_cast<aiVector3D *>(&node->translation.value));
|
||||||
|
|
||||||
|
|
|
@ -78,6 +78,7 @@ OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
||||||
#include <assimp/Profiler.h>
|
#include <assimp/Profiler.h>
|
||||||
#include <assimp/commonMetaData.h>
|
#include <assimp/commonMetaData.h>
|
||||||
|
|
||||||
|
#include <exception>
|
||||||
#include <set>
|
#include <set>
|
||||||
#include <memory>
|
#include <memory>
|
||||||
#include <cctype>
|
#include <cctype>
|
||||||
|
|
|
@ -44,6 +44,7 @@ OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
||||||
#ifndef INCLUDED_AI_IMPORTER_H
|
#ifndef INCLUDED_AI_IMPORTER_H
|
||||||
#define INCLUDED_AI_IMPORTER_H
|
#define INCLUDED_AI_IMPORTER_H
|
||||||
|
|
||||||
|
#include <exception>
|
||||||
#include <map>
|
#include <map>
|
||||||
#include <vector>
|
#include <vector>
|
||||||
#include <string>
|
#include <string>
|
||||||
|
|
|
@ -59,6 +59,8 @@ OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
||||||
// Public ASSIMP data structures
|
// Public ASSIMP data structures
|
||||||
#include <assimp/types.h>
|
#include <assimp/types.h>
|
||||||
|
|
||||||
|
#include <exception>
|
||||||
|
|
||||||
namespace Assimp {
|
namespace Assimp {
|
||||||
// =======================================================================
|
// =======================================================================
|
||||||
// Public interface to Assimp
|
// Public interface to Assimp
|
||||||
|
|
|
@ -247,14 +247,14 @@ typedef double ai_real;
|
||||||
typedef signed long long int ai_int;
|
typedef signed long long int ai_int;
|
||||||
typedef unsigned long long int ai_uint;
|
typedef unsigned long long int ai_uint;
|
||||||
#ifndef ASSIMP_AI_REAL_TEXT_PRECISION
|
#ifndef ASSIMP_AI_REAL_TEXT_PRECISION
|
||||||
#define ASSIMP_AI_REAL_TEXT_PRECISION 16
|
#define ASSIMP_AI_REAL_TEXT_PRECISION 17
|
||||||
#endif // ASSIMP_AI_REAL_TEXT_PRECISION
|
#endif // ASSIMP_AI_REAL_TEXT_PRECISION
|
||||||
#else // ASSIMP_DOUBLE_PRECISION
|
#else // ASSIMP_DOUBLE_PRECISION
|
||||||
typedef float ai_real;
|
typedef float ai_real;
|
||||||
typedef signed int ai_int;
|
typedef signed int ai_int;
|
||||||
typedef unsigned int ai_uint;
|
typedef unsigned int ai_uint;
|
||||||
#ifndef ASSIMP_AI_REAL_TEXT_PRECISION
|
#ifndef ASSIMP_AI_REAL_TEXT_PRECISION
|
||||||
#define ASSIMP_AI_REAL_TEXT_PRECISION 8
|
#define ASSIMP_AI_REAL_TEXT_PRECISION 9
|
||||||
#endif // ASSIMP_AI_REAL_TEXT_PRECISION
|
#endif // ASSIMP_AI_REAL_TEXT_PRECISION
|
||||||
#endif // ASSIMP_DOUBLE_PRECISION
|
#endif // ASSIMP_DOUBLE_PRECISION
|
||||||
|
|
||||||
|
|
|
@ -1,4 +0,0 @@
|
||||||
code/*.{%{cpp}}
|
|
||||||
contrib/**/*.{%{cpp}}
|
|
||||||
include/**/*.{%{cpp}}
|
|
||||||
test/**/*.{%{cpp}}
|
|
|
@ -1,177 +0,0 @@
|
||||||
#!/usr/bin/env ruby
|
|
||||||
|
|
||||||
#
|
|
||||||
# mtime_cache
|
|
||||||
# Copyright (c) 2016 Borislav Stanimirov
|
|
||||||
#
|
|
||||||
# Permission is hereby granted, free of charge, to any person obtaining a copy
|
|
||||||
# of this software and associated documentation files (the "Software"), to
|
|
||||||
# deal in the Software without restriction, including without limitation the
|
|
||||||
# rights to use, copy, modify, merge, publish, distribute, sublicense, and/or
|
|
||||||
# sell copies of the Software, and to permit persons to whom the Software is
|
|
||||||
# furnished to do so, subject to the following conditions:
|
|
||||||
#
|
|
||||||
# The above copyright notice and this permission notice shall be included in
|
|
||||||
# all copies or substantial portions of the Software.
|
|
||||||
#
|
|
||||||
# THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
|
|
||||||
# IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
|
|
||||||
# FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
|
|
||||||
# AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
|
|
||||||
# LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
|
|
||||||
# FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS
|
|
||||||
# IN THE SOFTWARE.
|
|
||||||
#
|
|
||||||
|
|
||||||
require 'digest/md5'
|
|
||||||
require 'json'
|
|
||||||
require 'fileutils'
|
|
||||||
|
|
||||||
VERSION = "1.0.2"
|
|
||||||
|
|
||||||
VERSION_TEXT = "mtime_cache v#{VERSION}"
|
|
||||||
|
|
||||||
USAGE = <<ENDUSAGE
|
|
||||||
|
|
||||||
Usage:
|
|
||||||
mtime_cache [<globs>] [-g globfile] [-d] [-q|V] [-c cache]
|
|
||||||
ENDUSAGE
|
|
||||||
|
|
||||||
HELP = <<ENDHELP
|
|
||||||
|
|
||||||
Traverse through globbed files, making a json cache based on their mtime.
|
|
||||||
If a cache exists, changes the mtime of existing unchanged (based on MD5
|
|
||||||
hash) files to the one in the cache.
|
|
||||||
|
|
||||||
Options:
|
|
||||||
|
|
||||||
globs Ruby-compatible glob strings (ex some/path/**/*.java)
|
|
||||||
A extension pattern is allowd in the form %{pattern}
|
|
||||||
(ex some/path/*.{%{pattern1},%{pattern2}})
|
|
||||||
The globs support the following patterns:
|
|
||||||
%{cpp} - common C++ extensions
|
|
||||||
|
|
||||||
-g, --globfile A file with list of globs to perform (one per line)
|
|
||||||
|
|
||||||
-?, -h, --help Show this help message.
|
|
||||||
-v, --version Show the version number (#{VERSION})
|
|
||||||
-q, --quiet Don't log anything to stdout
|
|
||||||
-V, --verbose Show extra logging
|
|
||||||
-d, --dryrun Don't change any files on the filesystem
|
|
||||||
-c, --cache Specify the cache file for input and output.
|
|
||||||
[Default is .mtime_cache.json]
|
|
||||||
|
|
||||||
ENDHELP
|
|
||||||
|
|
||||||
param_arg = nil
|
|
||||||
ARGS = { :cache => '.mtime_cache.json', :globs => [] }
|
|
||||||
|
|
||||||
ARGV.each do |arg|
|
|
||||||
case arg
|
|
||||||
when '-g', '--globfile' then param_arg = :globfile
|
|
||||||
when '-h', '-?', '--help' then ARGS[:help] = true
|
|
||||||
when '-v', '--version' then ARGS[:ver] = true
|
|
||||||
when '-q', '--quiet' then ARGS[:quiet] = true
|
|
||||||
when '-V', '--verbose' then ARGS[:verbose] = true
|
|
||||||
when '-d', '--dryrun' then ARGS[:dry] = true
|
|
||||||
when '-c', '--cache' then param_arg = :cache
|
|
||||||
else
|
|
||||||
if param_arg
|
|
||||||
ARGS[param_arg] = arg
|
|
||||||
param_arg = nil
|
|
||||||
else
|
|
||||||
ARGS[:globs] << arg
|
|
||||||
end
|
|
||||||
end
|
|
||||||
end
|
|
||||||
|
|
||||||
def log(text, level = 0)
|
|
||||||
return if ARGS[:quiet]
|
|
||||||
return if level > 0 && !ARGS[:verbose]
|
|
||||||
puts text
|
|
||||||
end
|
|
||||||
|
|
||||||
if ARGS[:ver] || ARGS[:help]
|
|
||||||
log VERSION_TEXT
|
|
||||||
exit if ARGS[:ver]
|
|
||||||
log USAGE
|
|
||||||
log HELP
|
|
||||||
exit
|
|
||||||
end
|
|
||||||
|
|
||||||
if ARGS[:globs].empty? && !ARGS[:globfile]
|
|
||||||
log 'Error: Missing globs'
|
|
||||||
log USAGE
|
|
||||||
exit 1
|
|
||||||
end
|
|
||||||
|
|
||||||
EXTENSION_PATTERNS = {
|
|
||||||
:cpp => "c,cc,cpp,cxx,h,hpp,hxx,inl,ipp,inc,ixx"
|
|
||||||
}
|
|
||||||
|
|
||||||
cache_file = ARGS[:cache]
|
|
||||||
|
|
||||||
cache = {}
|
|
||||||
|
|
||||||
if File.file?(cache_file)
|
|
||||||
log "Found #{cache_file}"
|
|
||||||
cache = JSON.parse(File.read(cache_file))
|
|
||||||
log "Read #{cache.length} entries"
|
|
||||||
else
|
|
||||||
log "#{cache_file} not found. A new one will be created"
|
|
||||||
end
|
|
||||||
|
|
||||||
globs = ARGS[:globs].map { |g| g % EXTENSION_PATTERNS }
|
|
||||||
|
|
||||||
globfile = ARGS[:globfile]
|
|
||||||
if globfile
|
|
||||||
File.open(globfile, 'r').each_line do |line|
|
|
||||||
line.strip!
|
|
||||||
next if line.empty?
|
|
||||||
globs << line % EXTENSION_PATTERNS
|
|
||||||
end
|
|
||||||
end
|
|
||||||
|
|
||||||
if globs.empty?
|
|
||||||
log 'Error: No globs in globfile'
|
|
||||||
log USAGE
|
|
||||||
exit 1
|
|
||||||
end
|
|
||||||
|
|
||||||
files = {}
|
|
||||||
num_changed = 0
|
|
||||||
|
|
||||||
globs.each do |glob|
|
|
||||||
Dir[glob].each do |file|
|
|
||||||
next if !File.file?(file)
|
|
||||||
|
|
||||||
mtime = File.mtime(file).to_i
|
|
||||||
hash = Digest::MD5.hexdigest(File.read(file))
|
|
||||||
|
|
||||||
cached = cache[file]
|
|
||||||
|
|
||||||
if cached && cached['hash'] == hash && cached['mtime'] < mtime
|
|
||||||
mtime = cached['mtime']
|
|
||||||
|
|
||||||
log "mtime_cache: changing mtime of #{file} to #{mtime}", 1
|
|
||||||
|
|
||||||
File.utime(File.atime(file), Time.at(mtime), file) if !ARGS[:dry]
|
|
||||||
num_changed += 1
|
|
||||||
else
|
|
||||||
log "mtime_cache: NOT changing mtime of #{file}", 1
|
|
||||||
end
|
|
||||||
|
|
||||||
files[file] = { 'mtime' => mtime, 'hash' => hash }
|
|
||||||
end
|
|
||||||
end
|
|
||||||
|
|
||||||
log "Changed mtime of #{num_changed} of #{files.length} files"
|
|
||||||
log "Writing #{cache_file}"
|
|
||||||
|
|
||||||
if !ARGS[:dry]
|
|
||||||
dirname = File.dirname(cache_file)
|
|
||||||
unless File.directory?(dirname)
|
|
||||||
FileUtils.mkdir_p(dirname)
|
|
||||||
end
|
|
||||||
File.open(cache_file, 'w').write(JSON.pretty_generate(files))
|
|
||||||
end
|
|
Loading…
Reference in New Issue