Hello ,
Hope you are doing great !

Please find the below mentioned job description and revert back with the
updated resume to *mahesh.gur...@solioscorp.com
<mahesh.gur...@solioscorp.com>*

Need Only - H1/GC/USC
If H1 Need Passport Number






*Position: Data Engineer with Python & AWSLocation: Dallas, TX - (Onsite
from Day 1 - Need Only Locals)Duration: Long Term ContractInterview Mode:
One Teams call, One Face to Face interviewData Engineer Responsibilities:*

   - Work with business stakeholders, Business Systems Analysts and
   Developers to ensure quality delivery of software.
   - Interact with key business functions to confirm data quality policies
   and governed attributes.
   - Follow quality management best practices and processes to bring
   consistency and completeness to integration service testing
   - Designing and managing the testing AWS environments of data workflows
   during development and deployment of data products
   - Provide assistance to the team in Test Estimation & Test Planning
   - Design, development of Reports and dashboards.
   - Analyzing and evaluating data sources, data volume, and business rules.
   - Proficiency with SQL, familiarity with Python, Scala, Athena, EMR,
   Redshift and AWS.
   - No SQL data and unstructured data experience.
   - Extensive experience in programming tools like Map Reduce to HIVEQL
   - Experience in data science platforms like SageMaker/Machine Learning
   Studio/ H2O.
   - Should be well versed with the Data flow and Test Strategy for Cloud/
   On Prem ETL Testing.
   - Interpret and analyses data from various source systems to support
   data integration and data reporting needs.
   - Experience in testing Database Application to validate source to
   destination data movement and transformation.
   - Work with team leads to prioritize business and information needs.
   - Develop complex SQL scripts (Primarily Advanced SQL) for Cloud ETL and
   On prem.
   - Develop and summarize Data Quality analysis and dashboards.
   - Knowledge of Data modeling and Data warehousing concepts with emphasis
   on Cloud/ On Prem ETL.
   - Execute testing of data analytic and data integration on time and
   within budget.
   - Work with team leads to prioritize business and information needs
   - Troubleshoot & determine best resolution for data issues and anomalies
   - Experience in Functional Testing, Regression Testing, System Testing,
   Integration Testing & End to End testing.
   - Has deep understanding of data architecture & data modeling best
   practices and guidelines for different data and analytic platforms

*Requirements:*

   - Extensive Experience in Data migration is a must (Teradata to Redshift
   preferred)
   - Extensive testing Experience with SQL/Unix/Linux scripting is a must
   - Extensive experience testing Cloud/On Prem ETL (e.g. Abinitio,
   Informatica, SSIS, Datastage, Alteryx, Glu)
   - Extensive experience DBMS like Oracle, Teradata, SQL Server, DB2,
   Redshift, Postgres and Sybase.
   - Extensive experience using Python scripting and AWS and Cloud
   Technologies.
   - Extensive experience using Athena, EMR, Redshift, AWS, and Cloud
   Technologies
   - Experienced in large-scale application development testing – Cloud/ On
   Prem Data warehouse, Data Lake, Data science
   - Experience with multi-year, large-scale projects
   - Expert technical skills with hands-on testing experience using SQL
   queries.
   - Extensive experience with both data migration and data transformation
   testing
   - Extensive experience DBMS like Oracle, Teradata, SQL Server, DB2,
   Redshift, Postgres and Sybase.
   - Extensive testing Experience with SQL/Unix/Linux.
   - Extensive experience testing Cloud/On Prem ETL (e.g. Abinitio,
   Informatica, SSIS, Datastage, Alteryx, Glu)
   - Extensive experience using Python scripting and AWS and Cloud
   Technologies.
   - Extensive experience using Athena, EMR , Redshift and AWS and Cloud
   Technologies
   - API/Rest Assured automation, building reusable frameworks, and good
   technical expertise/acumen
   - Java/Java Script - Implement core Java, Integration, Core Java and API.
   - Functional/UI/ Selenium - BDD/Cucumber, Specflow, Data
   Validation/Kafka, BigData, also automation experience using Cypress.
   - AWS/Cloud - Jenkins/ Gitlab/ EC2 machine, S3 and building Jenkins and
   CI/CD pipelines, SouceLabs.
   - API/Rest API - Rest API and Micro Services using JSON, SoapUI
   - Extensive experience in DevOps/Data Ops space.
   - Strong experience in working with DevOps and build pipelines.
   - Strong experience of AWS data services including Redshift, Glue,
   Kinesis, Kafka (MSK) and EMR/ Spark, Sage Maker etc…
   - Experience with technologies like Kubeflow, EKS, Docker
   - Extensive experience using No SQL data and unstructured data
   experience like MongoDB, Cassandra, Redis, ZooKeeper.
   - Extensive experience in Map reduce using tools like Hadoop, Hive, Pig,
   Kafka, S4, Map R.
   - Experience using Jenkins and Gitlab
   - Experience using both Waterfall and Agile methodologies.
   - Experience in testing storage tools like S3, HDFS
   - Experience with one or more industry-standard defect or Test Case
   management Tools
   - Great communication skills (regularly interacts with cross functional
   team members)


-- 
*Warm Regards,*

*Mahesh G*
Sr. US IT Recruitment Lead
mahesh.gur...@solioscorp.com
www.solioscorp.com

-- 
You received this message because you are subscribed to the Google Groups 
"VB.NET 2003 Group" group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to vbnet2003group+unsubscr...@googlegroups.com.
To view this discussion visit 
https://groups.google.com/d/msgid/vbnet2003group/CAAqQ8G3TrB_VKa-f_L8vEkY6nAao8xEAhiCMzWbrZwHrqOKryQ%40mail.gmail.com.

Reply via email to