September 6, 2018

Sreekanth B

Syntel Pentaho Recently Asked Interview Questions Answers

Explain The Use Of Pentaho Reporting.?

Pentaho reporting enables businesses to create structured and informative reports to easily access, format and deliver meaningful and important information to clients and customers. They also help business users to analyze and track consumer behavior for the specific time and functionality, thereby directing them towards the right success path.

What Is Pentaho Data Mining?

Pentaho Data Mining refers to the Weka Project, which consists of a detailed tool set for machine learning and data mining. Weka is open source software for extracting large sers of information about users, clients and businesses. It is built on Java programming.

Is Data Integration And Etl Programming Same?

No. Data Integration refers to passing of data from one type of systems to other within the same application. On the contrary, ETL is used to extract and access data from different sources. And transform it into other objects and tables.

Explain Hierarchy Flattening.?

It is just the construction of parent child relationships in a database. Hierarchy Flattening uses both horizontal and vertical formats, which enables easy and trouble-free identification of sub elements. It further allows users to understand and read the main hierarchy of BI and includes Parent column, Child Column, Parent attributes and Child attributes.

Syntel Pentaho Recently Asked Interview Questions Answers
Syntel Pentaho Recently Asked Interview Questions Answers

Define Pentaho Bi Project?

The Pentaho BI Project is an current effort by the Open Source communal to provide groups with best-in-class solutions for their initiative Business Intelligence (BI) needs.

What Major Applications Comprises Of Pentaho Bi Project?

The Pentaho BI Project encompasses the following major application areas:

Business Intelligence Platform
Data Mining
Reporting
Dashboards
Business Intelligence Platform

Which Platform Benefits From The Pentaho Bi Project?

Java developers who generally use project components to rapidly assemble custom BI solutions
ISVs who can improve the value and ability of their solutions by embedding BI functionality
End-Users who can quickly deploy packaged BI solutions which are either modest or greater to traditional commercial offerings at a dramatically lower cost.

Explain Pentaho?

It addresses the blockades that block the organization’s ability to get value from all our data. Pentaho is discovered to ensure that each member of our team from developers to business users can easily convert data into value.

How Do You Duplicate A Field In A Row In A Transformation?

Several solutions exist:

Use a “Select Values” step renaming a field while selecting also the original one. The result will be that the original field will be duplicated to another name.

It will look as follows:

This will duplicate fieldA to fieldB and fieldC.

Use a calculator step and use e.g. The NLV(A,B) operation as follows:

This will have the same effect as the first solution: 3 fields in the output which are copies of each other: fieldA, fieldB, and fieldC.

Use a JavaScript step to copy the field:

This will have the same effect as the previous solutions: 3 fields in the output which are copies of each other: fieldA, fieldB, and fieldC.

Why Can’t I Duplicate Fieldnames In A Single Row?

You can’t. PDI will complain in most of the cases if you have duplicate fieldnames. Before PDI v2.5.0 you were able to force duplicate fields, but also only the first value of the duplicate fields could ever be used.

I’ve Got A Transformation That Doesn’t Run Fast Enough, But It Is Hard To Tell In What Order To Optimize The Steps. What Should I Do?

Transformations stream data through their steps:

That means that the slowest step is going to determine the speed of a transformation.
So you optimize the slowest steps first. How can you tell which step is the slowest: look at the size of the input buffer in the log view.
In the latest 3.1.0-M1 nightly build you will also find a graphical overview of this: HTTP://WWW.IBRIDGE.BE/?P=92
(the “graph” button at the bottom of the log view will show the details).
A slow step will have consistently large input buffer sizes. A fast step will consistently have low input buffer sizes.

We Will Be Using Pdi Integrated In A Web Application Deployed On An Application Server. We’ve Created A Jndi Datasource In Our Application Server. Of Course Spoon Doesn’t Run In The Context Of The Application Server, So How Can We Use The Jndi Data Source In Pdi?

If you look in the PDI main directory you will see a sub-directory “simple-jndi”, which contains a file called “jdbc.properties”. You should change this file so that the JNDI information matches the one you use in your application server.

After that you set in the connection tab of Spoon the “Method of access” to JNDI, the “Connection type” to the type of database you’re using. And “Connection name” to the name of the JDNI datasource (as used in “jdbc.properties”).

Mention Major Features Of Pentaho?

Direct Analytics on MongoDB: It authorizes business analysts and IT to access, analyze, and visualize MongoDB data.

Science Pack: Pentaho’s Data Science Pack operationalizes analytical modeling and machine learning while allowing data scientists and developers to unburden the labor of data preparation to Pentaho Data Integration.

Full YARN Support for Hadoop: Pentaho’s YARN mixing enables organizations to exploit the full computing power of Hadoop while leveraging existing skillsets and technology investments.

Define Tuple?

Finite ordered list of elements is called as tuple.

What Kind Of Data, Cube Contain?

The Cube will contain the following data:

3 Fact fields: Sales, Costs and Discounts
Time Dimension: with the following hierarchy: Year, Quarter and Month

2 Customer Dimensions: one with location (Region, Country) and the other with Customer Group and Customer Name
Product Dimension: containing a Product Name

Differentiate Between Transformations And Jobs?

Transformations is moving and transforming rows from source to target.
Jobs are more about high level flow control.

Subscribe to get more Posts :