The Most Recent Guide To C2090-102 Questions
Proper study guides for Latest IBM IBM Big Data Architect certified begins with IBM C2090-102 preparation products which designed to deliver the Real C2090-102 questions by making you pass the C2090-102 test at your first time. Try the free C2090-102 demo right now.
Online C2090-102 free questions and answers of New Version:
NEW QUESTION 1
A large application vendor wants to port their existing distributed applications to run on Hadoop. In order to be competitive they need to provide monitoring and keep the size of the monitored applications consistent with the configuration. This implies the ability to deploy a replacement, for example, for any failed components. Which of the following would be a workable solution?
- A. Nagios with YARN
- B. Slider with YARN
- C. Oozie with Lucene
- D. OPTIM Performance Manager
Answer: B
NEW QUESTION 2
An upstream Oil and Gas Producer needs to optimize the performance of its assets. It needs to calculate Key Performance Indicators for flow rate sensors deployed to monitor the output flow and temperature and pressure for multiple pipelines in an oil field with hundreds of wells. Which of the following would you recommend to meet these requirements?
- A. Datastage ETL jobs should be created
- B. Infosphere Streams should be used
- C. Cognos reports would suffice here
- D. Hadoop based datastorage engine should be used
Answer: B
NEW QUESTION 3
Company K is designing their Big Data system. In their enterprise, they anticipate every 9 months there will be a big spike of new data on the order of multiple TB. Their company policy also dictates that data older than one year will be archived with a major clean up every 5 years. Cost is also a big issue. Which of the following provides the best design for these requirements?
- A. Estimate the peak volume over a 5 year period and set up a Hadoop system with commodity HW and storage to accommodate that volume
- B. Estimate the peak volume over a 3 year period and set up a Hadoop system with NAS to accommodate the expected volume
- C. Use Cloud elasticity capabilities to handle the peak and valley data volume
- D. Use SAN storage with compression to handle the peak and valley data volume
Answer: A
NEW QUESTION 4
Which one of the following statements about Big SQL is TRUE?
- A. Big SQL doesn’t need any secondary indices to access HBase tables
- B. Big SQL processes queries locally either on disk or in memory
- C. Big SQL supports updates in Hive.
- D. Executing Big SQL queries through MapReduce framework would always be a better choice
Answer: C
Explanation:
References:
http://www.ibm.com/developerworks/library/bd-bigsql/
NEW QUESTION 5
Which of the following is a consideration in sizing an active archive Hadoop infrastructure?
- A. replication factor within Hadoop
- B. Reporting requirements
- C. velocity or rate at which data is being generated
- D. veracity or trustworthiness of the data
Answer: B
Explanation:
Reference:
http://www.ibm.com/developerworks/library/ba-augment-data-warehouse3/
NEW QUESTION 6
You have a need for Storm real time processing and you realize that your Storm processing is detrimental to the timely execution of your MapReduce batch jobs. Which of the following would be your best course of action?
- A. Implement a Storm-YARN integration to facilitate the management of elastic workloads
- B. Implement the Oozie 2.0 framework optimized for elastic workload management
- C. Use a combination of Flume and Oozie 2.0 to enable the Flume built-in elastic flow automation
- D. Implement Apache ACE 2.0 for Storm
Answer: C
Explanation:
Reference:
http://www.ibm.com/developerworks/library/bd-ooziehadoop/
NEW QUESTION 7
Faced with a wide area network implementation, you have a need for asynchronous remote updates. Which one of the following would best address this use case?
- A. GPFS Active File Management allows data access and modifications even when remote storage cluster is unavailable
- B. HDFS Cluster rebalancing is compatible with data rebalancing scheme
- C. A scheme might automatically move data from one DataNode to another if the free space on a DataNode falls below a certain threshold
- D. GPFS File clones can be created from a regular file or a file in a snapshot using the mmclone command
- E. HDFS NameNode The NameNode keeps an image of the entire file systemnamespace and file Blockmap in memor
- F. This key metadata item is designed to be compact, such that a NameNode with 4 GB of RAM is plenty to support a huge number of files and directories
Answer: C
Explanation:
Reference:
http://www-01.ibm.com/support/knowledgecenter/STXKQY_4.1.1/com.ibm.spectrum.scale.v4 r11.adv.d oc/bl1adv_clones.htm
NEW QUESTION 8
A quality requirement or constraint that a system (or some part of a system) must satisfy is referred to by which of the following?
- A. Define Objective
- B. Non-functional Requirement
- C. Use Case Definition
- D. Service Level Agreement
Answer: D
NEW QUESTION 9
The YARN High Availability feature adds redundancy in the form of an Active/Standby. Which of the following will pair to remove this otherwise single point of failure?
- A. JobTracker
- B. Data Node
- C. Management Node
- D. Resource Manager
Answer: D
Explanation:
References:
http://hadoop.apache.org/docs/current/hadoop-yarn/hadoop-yarn-site/ResourceManagerHA.html
NEW QUESTION 10
In computer science, which of the following represents a set of properties that guarantee that database transactions are processed reliably?
- A. Foreign Key
- B. Primary Key
- C. Buffer Cache
- D. ACID
Answer: D
Explanation:
References:
https://en.wikipedia.org/wiki/ACID
NEW QUESTION 11
Defining your need to enrich existing customer data you realize that you need to process large quantities of Geospatial data and output to your data warehouse in a standard GeoJSON format. Which of the following would provide a business analyst with the desired output?
- A. Big SQL
- B. BigSheets
- C. Hive queries
- D. Text Analytics
Answer: C
NEW QUESTION 12
Which of the following statements regarding SPSS is TRUE?
- A. SPSS software provides a security framework
- B. SPSS analytics are primarily accessed through a scripting language
- C. SPSS models can only be applied to data at rest
- D. SPSS can directly use BigR syntax such as “bigr.list”
Answer: A
Explanation:
Reference:
http://www.ibmbigdatahub.com/blog/7-really-good-reasons-partner-spss-analytics-
r
NEW QUESTION 13
A bank wants to build a system that tracks all ATM and online transactions in real- time. They want to build a personalized model of their customer’s financial activity by incorporating enterprise data as well as social media data. The system must be able to learn and adapt over a period of time. These personalized models will be used for real time promotions as well as for any fraud or crime detections. Given these requirements, which of the following would recommend?
- A. Spark
- B. Hadoop
- C. Cloudand
- D. Netezza
Answer: D
NEW QUESTION 14
You are consulting with a company to design a new Big Data System. The CIO insists that all components must be open source. The key requirement is to provide text search on their data. For audit purposes the infrastructure must also be monitored and an alert system in place. To meet these requirements, which of the following would you suggest?
- A. HBase together with OPTIM performance Manager
- B. Text Analytics with OPTIM performance Manager
- C. Nagios with Lucene
- D. Oozie with Lucene
Answer: C
Explanation:
References: https://www.nagios.org/ https://www.ibm.com/support/knowledgecenter/SSPT3X_2.1.2/ com.ibm.swg.im.infosphere.biginsights.product.doc/doc/bi_addtl_technologies.ht ml
NEW QUESTION 15
Which of the following statements is TRUE regarding cloud applications?
- A. Migrating a legacy application to the cloud is a simple solution to drive down cost
- B. Architecting and deploying a scalable cloud application requires a private cloud implementation
- C. To be truly elastic, a cloud application must be tightly coupled to the physical cloud environment
- D. Leveraging a private v
- E. public cloud may result in sacrificing some of the core advantages of cloud computing
Answer: B
NEW QUESTION 16
A telco company is struggling to retain their high value customers.The marketing team has access to the billing data and is using it for customer segmentation. They would like to tap on network and social data to understand customer behavior before crafting any special offer. How best can they achieve this task?
- A. By loading both network and social data in the current Enterprise Data Warehouse, then run analytics
- B. By loading network and social data in BigInsights for exploration then moving resulting data to Enterprise Data Warehouse, and merging with billing data for analytics
- C. By loading network, social, and billing data in BigInsights and using available analytics tools
- D. By creating a dedicated data mart in their current Enterprise Data Warehouse
Answer: B
Explanation:
Reference:
https://ibmdatawarehousing.wordpress.com/category/ibm-data-warehouse/
NEW QUESTION 17
Which of the following statements is TRUE regarding cloud based solutions?
- A. In a Platform as a Service Cloud deployment, the customer chooses the operating system they want to use
- B. Automated recovery from hardware or network failures is not possible in a public cloud implementation, onlyin a private clouds
- C. There are benefits to use the cloud even for small-scale applications
- D. Using firewalls to create network boundaries is sufficient for ensuring cloud security
Answer: C
Explanation:
References:
http://www.ibm.com/developerworks/cloud/library/cl-cloudappdevelop/
NEW QUESTION 18
What is the most important aspect of data center disaster recovery?
- A. A complete damage assessment
- B. Control of critical assets
- C. Restoration of business functions
- D. Protection of individual life
Answer: D
NEW QUESTION 19
Which of the following requirements would NOT be effectively addressed by a NoSQL data store?
- A. Scalability
- B. Reporting
- C. Sparse data
- D. Batch processing
Answer: D
NEW QUESTION 20
Which of the following can NOT be performed by the Open Data Platform (ODP)?
- A. Provide development and governance of upstream projects
- B. Contribute to the Apache Software Foundation projects
- C. Produce a set of tools and methods that enable members to create and test differentiated offerings based on the ODP core
- D. Accelerate the delivery of Big Data solutions by providing a well-defined core consisting of all the components of the Apache Software Foundation projects
Answer: C
NEW QUESTION 21
A manufacturing company has decided they need to capture and analyze the log files of their software automation system. Their business users are still trying to define the use cases but would want to start capturing as they have had frequent outages. Given this, which of the following is the best software design recommendation?
- A. ETL tools and a Data Warehouse
- B. Flume and Hadoop
- C. Pure Data for Analytics and Optim
- D. Streams and BigInsights
Answer: D
NEW QUESTION 22
SPSS and BigInsights integration is enabled by which of the following?
- A. SPSS Analytic Server
- B. InfoSphere DataStage
- C. SPSS Modeler
- D. SPSS Collaboration & Deployment Services
Answer: A
Explanation:
References:
http://www.ibm.com/developerworks/library/bd-spss/
NEW QUESTION 23
The AQL query language is the easiest and most flexible tool to pull structured output from which of the following?
- A. Hive data structures
- B. Unstructured text
- C. Hbase schemas
- D. JDBC connected relational data marts
Answer: A
Explanation:
Reference:
http://www.ibm.com/developerworks/library/bd-sqltohadoop2/
NEW QUESTION 24
A major telecommunication company has millions of customers. Most of their customers are prepaid. Being prepaid customers, they can very easily switch to other vendors. The last four to six months, this company has lost quite a good number of customers to competition. They intend to build a system that can provide them with insight into the customer’s social network (e.g. who is the influencer and who is the follower). They also want the ability to monitor the voice
and data usage patterns in real time and they want the system to be trained over time to predict possible dissatisfactions. Given this scenario, which one of the following would you recommend?
- A. Hadoop
- B. Spark
- C. Cloudant
- D. Netezza
Answer: B
NEW QUESTION 25
You have implemented a large Hadoop MapReduce cluster and the applications and users are multiplying. You are now faced with requests for interactive and streaming data applications while you still need to support the original MapReduce batch Jobs. Select the best option for continued support and performance.
- A. Just add several data nodes as Hadoop clusters are designed to scale-up easily
- B. Keep your original cluster configuration, all that is needed is re-optimizing the Oozie- workflow management
- C. Implement Yarn to decouple MapReduce and resource management
- D. Implement Apache Cassandra to automatically optimize multi-tenancy workloads
Answer: D
NEW QUESTION 26
......
Recommend!! Get the Full C2090-102 dumps in VCE and PDF From DumpSolutions.com, Welcome to Download: https://www.dumpsolutions.com/C2090-102-dumps/ (New 110 Q&As Version)