Generating Output

  • Generating Output

    Overview Once you have built a recipe, you are ready to apply that recipe to the source data and generate an output. In Wrangler Desktop, select Generate Results and choose the output format you would like to generate: In Wrangler Enterprise, select Run Job, choose an environment to run on, and select a publishing action and location: Validating your results When you have generated your re…

  • FAQ: How does Trifacta security integrate with Hadoop?

    Trifacta can interact with the Hadoop cluster using two authentication modes: System mode and User Impersonation mode.In System mode, a single 'trifacta' system account performs actions in Hadoop. This account is given a Kerberos keytab for the cluster. The Trifacta server uses the keytab to authenticate using Kerberos delegation tokens to perform Hadoop actions, such as access WebHDFS o…

  • HOW TO: Change Output Home Directory

    1. Click the User Preference menu from the tool bar. 2. Select your name from the drop-down. 3. Update the path of your desired Output Home Directory or select 'Browse'.We recommend creating a new directory or using these default locations:For OS X: /Users/<username>/trifactaFor Windows: C:\Users\<username>\trifacta 4. Click 'Save'. NOTE: To access datas…

  • HOW TO: Generate Results or Run a Job

    After you have created your transformation steps and are happy with how they are applied to the sample in the Transformer page, you can apply the steps to your entire dataset. Follow the steps below: Click Run Job or Generate Results (Trifacta Wrangler):In the dialog, select the running environment: Trifacta Server for small files (<100MB) or Hadoop for larger files. For Options,…

  • FAQ: Which type of Job do I launch?

    Overview When interacting with Trifacta web application, you are operating and building transforms on a sample of your dataset. When you are happy with your transforms, you can run Trifacta on your entire dataset. Note: Trifacta will automatically suggest the best execution mode depending on your data.  1. Click on the blue Run Job button in the upper right.  2. Select the output fo…

  • FAQ: Troubleshooting a Hadoop Job Failure

    You can use the Trifacta and Hadoop job logs to troubleshoot when your Trifacta job fails to run at scale on Hadoop. The Trifacta logs are located under your root Trifacta installation in the “logs” directory. This directory contains the following logs:webapp.log. This log monitors when interaction with the Trifacta web application. You will be able to see issues related to jobs running locall…

  • HOW TO: View Results

    The Job Results page displays details about a finished job. This page contains useful information regarding the execution of the job, such as: a Data Quality Bar, which indicates data conformity, statistics about the values in each column, and histograms relating to each column. This page can be accessed in two ways: Click the View Results button from the Datasets tab, under the Jobs …

  • DEFINITION: Results Card

    You can use the Results Cards to view all of the result files that you have generated for a given dataset.The following table explains the elements shown on each Results Card: IconDescription View the summary profile for your result file. Open the result file. This file has been saved to your local machine. View the script that was used to generate the results. Shows the percentage of …

  • HOW TO: Download and Publish Results

    1.  From the Workspace view, click the name of the dataset for which you want to download the results of a job. In the following image, we want to download the job results for the dataset named "BANK_WEBLOGS": 2. The Dataset Details page opens.3. Click the down arrow on the Job Results Card, as shown in the following image: 4. The Export Results pop-up window opens. Click the nam…

  • FAQ: What file permissions does Trifacta set on job result files written to HDFS?

    The file permissions that Trifacta sets for job result files depend on two factors:How user accounts are configured on HDFS.How Trifacta is configured to interface with HDFS.See the following article for details on how Trifacta interfaces with Hadoop security:FAQ: How does Trifacta security integrate with Hadoop?If your Trifacta installation interacts with Hadoop through User Impersonation mode, t…

  • HOW TO: Download TDE (Tableau Data Extract) Results

    If the end results of your Trifacta datasets are destined to be used in Tableau, it’s very easy to output a TDE and there are a couple of ways to accomplish this. Read more about Tableau's TDE here.NOTE: Be sure to check the box TDE under the 'Formats' section when you Run/Generate Results. Method 1: Go to the Jobs/Results page. …

  • FAQ: What is a Job?

    In the Trifacta Data Transformation Platform, a "job" refers to an execution of a transformation script on an entire file (or files) of data. Pointers to historic jobs are available from the Jobs tab. From each job, the user can see the script that was executed, the profiling feedback that was generated, the location of the full output file in HDFS, and the user can download samples of t…