Big Data Job Description
Big Data Duties & Responsibilities
To write an effective big data job description, begin by listing detailed duties, responsibilities and expectations. We have included big data job description templates that you can modify and use.
Sample responsibilities for this position include:
Big Data Qualifications
Qualifications for a job description may include education, certification, and experience.
Licensing or Certifications for Big Data
List any licenses or certifications required by the position: AWS, GCP, SNS, SQS, API, S3, COMPTIA, PKI
Education for Big Data
Typically a job would require a certain level of education.
Employers hiring for the big data job most commonly would prefer for their future employee to have a relevant degree such as Bachelor's and Master's Degree in Computer Science, Engineering, Technical, Information Technology, Information Systems, Mathematics, Statistics, Math, Computer Engineering, Business
Skills for Big Data
Desired skills for big data include:
Desired experience for big data includes:
Big Data Examples
Big Data Job Description
- Define data standards for use across the ERS technology stack
- Assist in the transformation of high volume complex data structures into relevant information that can be used by Data Scientists, BI Analysts, and detection systems
- Work closely with product planning and business stakeholders to help define the strategy to best meet our customers’ needs
- Become an expert on system architecture, functionality, and data flows operations processes and tools
- Evangelize ERS Data Engineering products to data consumers
- Evaluate and make recommendations on changes in upstream technologies and assess impacts to data
- Effectively manage trade-offs between effort/schedule and customer benefits
- Collaborate with Engineers, Data Scientists, System Architects, Product Managers, Analysts, and Production Support Analysts to implement successful solutions
- Understand and manage customer requirements (delivery and environmental factors) to ensure contract understanding and acceptance for deliverables, cost, contract type, timeline, risk and assumptions, and customer satisfaction
- Ensure delivery methodology is understood by team from project initiation to project closeout
- Strong experience with the Atlassian JIRA, Jenkins, Stash and Confluence
- Must possess abilities in Microsoft Excel, PowerPoint, and Word
- 5-10 years as a power user of traditional BI tools – Business Objects, Microstrategy, Cognos
- 3-5 years of experience on leading analytic technologies – Tableau, Platfora, Mixpanel, Qlik
- Hungry to learn new technology – Big Data, Hadoop, A deep desire to be more techie
- Deep understanding of Apache Hadoop 1/2 and the Hadoop ecosystem
Big Data Job Description
- Mentor, manage and develop product/project team members
- Build and deliver compelling customer focused demonstrations
- Identify and implement areas for continuous process improvement
- Research, experiment and prototype new software technologies and concepts to improve our data lake and data model
- Ensure best practices for processing, cleansing, and verifying the integrity of data used for analysis and learning
- Contribute to data mining techniques and applications using state-of-the-art methods
- Working knowledge of Tensorflow or H20 or Scikit or related framework or libraries, along with discrete math concepts
- Collaborating with internal business clients to identify and develop enterprise scale analytic algorithms on large data sets
- Acquiring, curating, and cleaning data from a wide variety sources, both public and private
- Developing and maintaining ETL pipelines and workflows
- Master degree in computer science or related fields
- Expertise in machine learning, graph analytics and text mining techniques, such as classification, regression, clustering, feature engineering, label propagation, PageRank, information extraction, topic modeling
- Kaggle player will be a plus
- Developing models to generate actionable insights that help inform those business decisions
- Making clear, data-driven, and actionable, recommendations to the business on new initiatives and areas for potential improvement
- Presenting data and recommendations in a well-structured, clear, format
Big Data Job Description
- Design and Develop big data solutions using Hadoop and Spark platform
- Selecting and integrating big data tools and frameworks to support the business’ analytics needs
- Provisioning and maintaining cloud based databases and cloud computing environments
- Continuing to grow knowledge of data engineering tools and processes
- Evaluate trends, characterize factors, develop predictive models, and identify optimization opportunities.Provide accurate, timely, relevant and quality analysis that translate data into sound organizational decisions.Research, design and prototype robust and scalable models based on machine learning, data mining, and statistical modeling to answer key business problems
- Define conformed zone LDM/PDM, SLA of the data pipeline RFT Data Services data platform
- Define LDM (Logical Data Model) and/or LAM (Logical Access Model) in the Risk Repository
- Define boundary sets for the data contents RFT plays as System of Records
- Work closely with data engineering/IT/dev teams to architect and engineer solutions that provide solutions around our analytic and big data platforms
- Create and maintain conceptual, logical and physical data models of data assets & resources
- Understanding (and explaining in clear terms) the limitations and nuances of the models
- Experience in software development/ engineering, including requirements analysis, software development, installation, integration, evaluation, enhancement, maintenance, testing, and problem diagnosis/resolution
- Knowledge of Linear Algebra, ability to communicate the results of analyses
- Experience with the data/development architecture in the field of big data specific to Hadoop Zoo
- Hadoop/MapR
- Expertise in big data technologies and principles including architecture, design, data acquisition, data delivery methodologies, and performance optimization in multi-terabyte environments
Big Data Job Description
- Work with alternative analytic data systems to incorporate them into the operational data flow for the Analytics Teams
- Engage with large functional efforts across multiple projects with minimal guidance including to provide guidance and education to other senior and intermediate contributors within team
- Ensure data flows and related information is being tracked, managed and documented
- Keep current with Big Data technologies in order to recommend best approaches in order to perform current and future work
- Solution - Deliver on-site technical engagements with partners and customers
- Working with RFT Data Services Data Architect/Modeler and up-stream technology partners to define conformed zone LDM/PDM, SLA of the data pipeline RFT Data Services data platform
- Working with RFT Data Services Data Architect/Modeler and down-stream technology partners to define LDM (Logical Data Model) and/or LAM (Logical Access Model) in the Risk Repository
- Ensuring development outputs are suitably integrated into release modules
- The support of the release-to-service process including security and accreditation processes
- Providing Technical advice and expertise on designs, solutions, tools, techniques and standards
- Extensive programming experience in Java, Python or Scala – we’re looking for a coding geek!
- Driven, proactive, and accustomed to working with aggressive project timelines and with a history of delivering high quality technical solutions to data-driven business challenges
- Proficiency in data mining and analysis with intermediate to advanced query skills (ex
- Ability to understand JSON & XML streaming messages and create logical designs for transforming data into queriable structures
- Demonstrated ownership of projects in a technical environment, enabling teams during the software development lifecycle
- Ability to facilitate communication between technical and business stakeholders to ensure functional requirements and technical implementation considerations are understood and accepted
Big Data Job Description
- Continuously improve software engineering practices by adapting innovative ways
- Works with the teams to plan approach to understand current state architecture
- Contributing data sources, technologies, interfaces, performance issues and system configurations
- Design and analyze data models, data flows, and integration patterns for structural deficiencies/soundness to build a robust and complete future state model
- Create/Review logical entities and define their attributes
- Design physical data models and generate DDLs
- Perform data normalization and denormalization where needed
- Define volumetric and Define access requirements
- Design and deliver data solutions to empower data migration initiatives, BI initiatives, dashboards development
- Differentiate and recommend different tools that can be used to solve business problems
- Experience with Hadoop environment (including Hive, Impala, Sqoop, Spark )
- Talent for working cross functionally to drive the implementation of new features resolve operational issues
- Ability to learn quickly and identify opportunities for improvement
- Bachelor degree in Computer Science, related technical field preferred, or equivalent work experience
- 7 + years Hands on Experience with distributed environments such as HDFS/Hadoop, Cloudera stack-Impala, Sqoop
- 5 + years of experience with Oracle and 2 + years with Impala