Docsity
Docsity

Prepare for your exams
Prepare for your exams

Study with the several resources on Docsity


Earn points to download
Earn points to download

Earn points by helping other students or get them with a premium plan


Guidelines and tips
Guidelines and tips

Palantir Data Engineering Certification Exam Questions And Answers, Exams of Data Acquisition

Palantir Data Engineering Certification Exam Questions And Answers

Typology: Exams

2024/2025

Available from 02/26/2025

Lebron.
Lebron. 🇺🇸

2.5K documents

1 / 56

Toggle sidebar

This page cannot be seen from the preview

Don't miss anything!

bg1
Palantir Data Engineering Certification
Exam Questions And Answers
2. Which of the following is the correct sequence of steps to
configure a direct connection in Foundry's managed SaaS platform?
configure a network policy → provision credentials → create the
source in data connection → configure network egress policy
create the source in data connection → configure a network policy
→ configure network egress policy → provision credentials
provision credentials → configure network egress policy → create
the source in data connection → configure a network policy
configure a network egress policy → provision credentials → create
the source in data connection → configure a network policy -
correct answer configure a network egress policy → provision
credentials → create the source in data connection → configure a
network policy
5. You are responsible for integrating data from an Azure storage
account into Foundry. To ensure optimal uptime and performance
without managing additional infrastructure, which connection
method should you configure?
Third-Party Sync Tool
Agent-based Connection
pf3
pf4
pf5
pf8
pf9
pfa
pfd
pfe
pff
pf12
pf13
pf14
pf15
pf16
pf17
pf18
pf19
pf1a
pf1b
pf1c
pf1d
pf1e
pf1f
pf20
pf21
pf22
pf23
pf24
pf25
pf26
pf27
pf28
pf29
pf2a
pf2b
pf2c
pf2d
pf2e
pf2f
pf30
pf31
pf32
pf33
pf34
pf35
pf36
pf37
pf38

Partial preview of the text

Download Palantir Data Engineering Certification Exam Questions And Answers and more Exams Data Acquisition in PDF only on Docsity!

Exam Questions And Answers

  1. Which of the following is the correct sequence of steps to configure a direct connection in Foundry's managed SaaS platform? configure a network policy → provision credentials → create the source in data connection → configure network egress policy create the source in data connection → configure a network policy → configure network egress policy → provision credentials provision credentials → configure network egress policy → create the source in data connection → configure a network policy configure a network egress policy → provision credentials → create the source in data connection → configure a network policy - correct answer ✅configure a network egress policy → provision credentials → create the source in data connection → configure a network policy
  2. You are responsible for integrating data from an Azure storage account into Foundry. To ensure optimal uptime and performance without managing additional infrastructure, which connection method should you configure? Third-Party Sync Tool Agent-based Connection

Exam Questions And Answers

Manual Network Tunneling Direct Connection - correct answer ✅Direct Connection

  1. What is the minimum recommended amount of RAM for a Foundry agent host? 12 GB 8 GB 32 GB 16 GB - correct answer ✅16 GB
  2. Which of the following are part of securing a Foundry agent host? Select two. Allow all inbound traffic to facilitate connectivity. Allow network traffic only from specific IPs. Open all ports for flexibility. Install antivirus software on the host.

Exam Questions And Answers

Change the branch of the dataset. Modify the export column type to match desired data types. Delete the original Fusion sheet without affecting the dataset. Use both sheet sync and table sync on the same Fusion sheet. Automatically merge changes from multiple Fusion sheets. Rename the synced dataset. - correct answer ✅Change the branch of the dataset. Modify the export column type to match desired data types. Rename the synced dataset.

  1. Which open data format is used by default for transformed data in Palantir AIP to ensure compatibility with existing data architectures? JSON Parquet CSV Avro - correct answer ✅Parquet

Exam Questions And Answers

  1. Which of the following are responsibilities of Action types in the Palantir Ontology? Select two. Provide object type polymorphism Define link types Capture data from operators Author business logic Orchestrate decision-making processes Define object properties - correct answer ✅Capture data from operators Orchestrate decision-making processes
  2. You are responsible for syncing a specific range of data from a Fusion spreadsheet to a dataset in Foundry to be used by Contour. After selecting the desired table range and initiating the sync, what must you ensure to avoid synchronization issues? Ensure that the dataset has Viewer permissions. Export the synced data as a CSV file immediately after syncing.

Exam Questions And Answers

Storing the unstructured data as binary blobs without parsing. Leveraging custom Python or Java code within the transform to handle parsing. Using built-in SQL functions to parse the data directly. - correct answer ✅Leveraging custom Python or Java code within the transform to handle parsing.

  1. Which role is required to configure network egress policies in Foundry's managed SaaS platform? Information Security Officer User Project Admin Data Pipeline Developer - correct answer ✅Information Security Officer
  2. Which of the following components enhance security interoperability within Palantir AIP? SAML integration for authentication

Exam Questions And Answers

Using internal scripts for authorization Role-based permissions Proprietary authentication systems Permissions managed through JSON files Integration with Active Directory - correct answer ✅SAML integration for authentication Role-based permissions Integration with Active Directory

  1. Which of the following practices are essential when implementing pipelines that back ontology objects and links in Foundry? Select two. Aligning pipeline logic with the ontology's entity and relationship definitions. Using only default transformation settings without customization. Avoiding documentation to keep the pipeline simple. Manually verifying each pipeline run for consistency. Ensuring that data transformations preserve the integrity of semantic relationships.

Exam Questions And Answers

Create detailed documentation outlining common issues and troubleshooting steps. Set up automated alerting for pipeline failures and performance issues.

  1. A data scientist wants to leverage their existing Jupyter notebooks within Palantir AIP for data analysis without switching to a different interface. Which feature of Palantir AIP should they utilize to achieve this? REST Interfaces Virtual Tables Palantir HyperAuto Pipelines Code Workspaces - correct answer ✅Code Workspaces
  2. What are the kinetic elements in the Palantir Ontology? Objects, Properties, Links Actions, Functions Semantics, Interfaces

Exam Questions And Answers

Object Types, Link Types - correct answer ✅Actions, Functions

  1. Which Linux operating system version is specifically recommended for hosting a Foundry agent? Ubuntu 18. Fedora 34 Debian 10 Red Hat Enterprise Linux 8 - correct answer ✅Red Hat Enterprise Linux 8
  2. What actions are performed when the ModelOutput.publish() method is called in Foundry's Code Repositories? Select two: It serializes the model using the ModelAdapter.save() method. It initializes the model adapter with the fresh model. It runs the model inference.

Exam Questions And Answers

chain as many expressions as possible for conciseness. use backslashes () for line breaks in chains. limit chains to a maximum of 5 statements. extract complex logic into separate functions. nest multiple chains within a single expression block. - correct answer ✅limit chains to a maximum of 5 statements. extract complex logic into separate functions.

  1. You need to inject a TransformContext into your Transform's compute function to access the current Spark session. How should you define the parameters of your compute function? def compute(context, input, output): def compute(input, output): def compute(input, output, ctx): def compute(ctx, input, output): - correct answer ✅def compute(ctx, input, output):
  2. You have a dataset in the Foundry filesystem that includes JPEG and PDF files, and you want to upload only the PDF files to a media

Exam Questions And Answers

set. Which parameter can you use in the put_dataset_files() method to achieve this? upload_specific_types=['pdf'] only_upload_matching_files=True filter_schema=True ignore_items_not_matching_schema=True - correct answer ✅ignore_items_not_matching_schema=True

  1. What is the first step to set up media sets in your Python transform in Foundry? Initialize media sets using the @ initialize_media_set decorator add a dependency on 'transforms-media' in your code repository create media sets directly in the Python code use the @ media_set_input decorator to specify media sets - correct answer ✅add a dependency on 'transforms-media' in your code repository

Exam Questions And Answers

incremental - correct answer ✅incremental

  1. In Foundry, which schema field type requires specifying both precision and scale parameters? ARRAY DECIMAL DATE STRING - correct answer ✅DECIMAL
  2. Which of the following are stages included in the condaPackRun task used in CI checks for a Python repository in Foundry? Select three. Run unit tests using PyTest Upload artifacts to a remote server Download and extract all packages in the solved environment Compile the Python source code

Exam Questions And Answers

Link packages into the environment Verify package contents - correct answer ✅Download and extract all packages in the solved environment Link packages into the environment Verify package contents

  1. Which of the following Python libraries is NOT recommended for training models in Foundry's Code Repositories? scikit-learn SparkML PyTorch TensorFlow - correct answer ✅SparkML
  2. Which of the following are recommended practices for refactoring complex logical operations in PySpark transformations? Chain multiple 'filter()' and 'withColumn()' calls in a single line

Exam Questions And Answers

  1. You are performing a left join between two DataFrames in PySpark, but realize that the right DataFrame may have multiple matches for some keys, leading to duplicate rows in the output. According to the style guide, what should you do to prevent this 'join explosion'? Use .dropDuplicates() after the join Switch to an inner join to avoid duplicates Ensure the join key in the right DataFrame is unique Use a right join instead of a left join - correct answer ✅Ensure the join key in the right DataFrame is unique
  2. Which of the following are considered bad practices when performing joins in PySpark? Using dataframe aliases to disambiguate column names. Dropping unnecessary columns after the join. Ensuring the key you join on is unique when performing left joins. Using right joins.

Exam Questions And Answers

Explicitly specifying the join type. Allowing expressions that duplicate columns in the output. - correct answer ✅Using right joins. Allowing expressions that duplicate columns in the output.

  1. When defining Transform logic level versioning (TLLV), which of the following factors are included in the default version string? Select three. The names of all input datasets All modules the Transform depends on The module where the Transform is defined Any project dependencies The runtime environment configuration All functions within the Transform - correct answer ✅All modules the Transform depends on The module where the Transform is defined Any project dependencies