Content Transfer extract failed - Extracted text could not be read from indexes. | Community
Skip to main content
New Participant
April 28, 2023
Solved

Content Transfer extract failed - Extracted text could not be read from indexes.

  • April 28, 2023
  • 3 replies
  • 1239 views

Hi All,

I'm using the aem-sdk-2023.3.11382.20230315T073850Z-230200 on the local environment as the author instance.

I want to transfer local content to the sandbox env and follow the Getting Started with Content Transfer Tool

But migration set extraction failed.

 

 

 

2023-04-28 14:27:27,142 [main] INFO o.a.j.o.p.i.d.DataStoreTextWriter - Using /Users/stevenl/aem-sdk/cloud-auchor/crx-quickstart/cloud-migration/extraction-Dev/tmp/1682663247059-0/store to store the extracted text content. Empty count 0, Error count 0 [27.572s][info][gc] GC(15) Pause Young (Normal) (G1 Evacuation Pause) 149M->82M(256M) 2.533ms 2023-04-28 14:27:27,156 [main] ERROR c.a.g.s.m.c.a.TextRenditionRepositoryTransformer - Extracted text could not be read from indexes. [27.597s][info][gc] GC(16) Pause Full (System.gc()) 85M->64M(227M) 11.032ms 2023-04-28 14:27:27,179 [main] INFO o.a.j.oak.segment.file.FileStore - TarMK closed: crx-quickstart/cloud-migration/extraction-Dev/tmp/1682663225326-0 2023-04-28 14:27:27,184 [main] INFO c.a.g.s.m.c.s.AzureBlobStoreFactory - Directory crx-quickstart/cloud-migration/extraction-Dev/tmp/1682663245146-0 deleted 2023-04-28 14:27:27,184 [main] ERROR c.a.granite.skyline.migrator.Main - Error in migration [27.612s][info][gc] GC(17) Pause Full (System.gc()) 65M->44M(160M) 9.487ms 2023-04-28 14:27:27,194 [main] INFO o.a.j.o.s.file.ReadOnlyFileStore - TarMK closed: /Users/stevenl/aem-sdk/cloud-auchor/crx-quickstart/repository/segmentstore 2023-04-28 14:27:27,998 [oak-ds-async-upload-thread-4] ERROR o.a.j.o.p.blob.UploadStagingCache - Error adding file to backend org.apache.jackrabbit.core.data.DataStoreException: Cannot write blob. identifier=1050-ce794b8515e4df0894898fabb7dd00e9bb37adc659a4d5925342aa73d461 at org.apache.jackrabbit.oak.blob.cloud.azure.blobstorage.AzureBlobStoreBackend.write(AzureBlobStoreBackend.java:328) at org.apache.jackrabbit.oak.plugins.blob.AbstractSharedCachingDataStore$2.write(AbstractSharedCachingDataStore.java:173) at org.apache.jackrabbit.oak.plugins.blob.UploadStagingCache$3.call(UploadStagingCache.java:367) at org.apache.jackrabbit.oak.plugins.blob.UploadStagingCache$3.call(UploadStagingCache.java:362) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128) at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628) at java.base/java.lang.Thread.run(Thread.java:829) Caused by: java.io.FileNotFoundException: /Users/stevenl/aem-sdk/cloud-auchor/crx-quickstart/cloud-migration/extraction-Dev/tmp/1682663245146-0/repository/datastore/upload/10/50/ce/1050ce794b8515e4df0894898fabb7dd00e9bb37adc659a4d5925342aa73d461 (No such file or directory) at java.base/java.io.FileInputStream.open0(Native Method) at java.base/java.io.FileInputStream.open(FileInputStream.java:219) at java.base/java.io.FileInputStream.<init>(FileInputStream.java:157) at org.apache.jackrabbit.oak.blob.cloud.azure.blobstorage.AzureBlobStoreBackend.write(AzureBlobStoreBackend.java:294) ... 7 common frames omitted

 

 

Best regards.

This post is no longer active and is closed to new replies. Need help? Start a new post to ask your question.
Best answer by joerghoh

Please check with the Adobe support team and report this issue.

3 replies

joerghoh
joerghohAccepted solution
Employee
May 1, 2023

Please check with the Adobe support team and report this issue.

New Participant
April 28, 2023

Hi @krati_garg 
Thank you for your reply.

I've executed the revision clean up and data consistency check.

The data consistency check show the below message.

[ref]1682677902543: https://drive.google.com/file/d/16jqKsDScTktVLKAeYiWkWCRJDyyzEXqv/view?usp=share_link

[consistency]1682677902543:  https://drive.google.com/file/d/12jmN48xBAdOPzLU7TrkzEZwwzyw90giD/view?usp=share_link

 

What do these errors mean?

 

 

java -jar oak-run-1.50.0.jar datastorecheck --consistency --ref --id --fds crx-quickstart/install/org.apache.jackrabbit.oak.plugins.blob.datastore.FileDataStore.config --repoHome crx-quickstart/repository --store crx-quickstart/repository/segmentstore --dump temp --verbose Apache Jackrabbit Oak 1.50.0 Starting dump of blob ids 0 blob ids found Finished in 0 seconds Starting dump of blob references by traversing 1256 blob references found Finished in 0 seconds Starting consistency check Consistency check found 1199 missing blobs Consistency check failure for the data store Finished in 0 seconds [consistency] - /Users/stevenl/aem-sdk/cloud-auchor/temp/[consistency]1682677902543 [ref] - /Users/stevenl/aem-sdk/cloud-auchor/temp/[ref]1682677902543

 

 

 

 

krati_garg
Employee
April 28, 2023

@aloneblood Please confirm if you have executed Revision Clean up and Data Consistency Check, before starting extraction. Please consider below best practices and guidelines to begin with Content Transfer via CTT
https://experienceleague.adobe.com/docs/experience-manager-cloud-service/content/migration-journey/cloud-migration/content-transfer-tool/guidelines-best-practices-content-transfer-tool.html?lang=en