Troubleshooting Tableau collector issues
Collector runtime and troubleshooting
The catalog collector may run in several seconds to many minutes depending on the size and complexity of the system being crawled.
If the catalog collector runs without issues, you should see no output on the terminal, but a new file that matching *.dwec.ttl should be in the directory you specified for the output.
If there was an issue connecting or running the catalog collector, there will be either a stack trace or a *.log file. Both of those can be sent to support to investigate if the errors are not clear.
A list of common issues and problems encountered when running the collectors is available here.
Issue 1: The collector is taking a long time to harvest from Tableau
Cause: The large size of a Tableau environment results in a long time to harvest metadata.
Solution: Increase the graphql page size to reduce the number of API calls.
Issue 2: Out of memory errors while running the collector
Cause: There is not enough memory allocated to the collector on the machine that the collector is running. Note that the collector may run for a long time even if there is sufficient memory due to the size of the source system.
Solution: Increase memory of the machine running the collector. Run docker system info to see the total memory available for Docker. The collector will use 80% of what is allocated to the container. You can run docker stats to see how much memory is used by the container when the collector runs.
Issue 3: Info message "Authentication to Tableau API failed, reauthenticating" observed in log file
Cause: The connection to Tableau API expired due to a long run.
Solution: No action is required as the collector re-authenticates automatically to Tableau.
Issue 4: Partial results error observed
The following error message is observed: Showing partial results. The request exceeded the ‘n’ node limit. Use pagination, additional filtering, or both in the query to adjust results.
Cause: When you increase the graphql page size, you may run into warnings messages in the logs due to nested queries.
Solution: Try a smaller page size or increase the max node limit. Increase the max node limit by setting metadata.query.limits.count which defaults to 20,000.