Job Title: Data Engineer (US Citizens Only) Location: RemoteWe are currently seeking candidates who meet the following qualifications: Key Responsibilities:
Develop and maintain robust, scalable, and efficient data pipelines that integrate data from multiple sources.
Ensure timely, accurate, and high-quality data delivery across various systems.
Optimize data workflows for better performance, cost-efficiency, and reliability.
Extract, transform, and load (ETL) data from a variety of sources, including internal and third-party systems.
Work closely with stakeholders to understand business needs and transform data into actionable insights.
Develop and maintain a variety of data integration processes using modern ETL tools and techniques.
Maintain and optimize data storage systems such as SQL databases, NoSQL databases, and data lakes.
Ensure high availability, security, and performance of data systems.
Monitor and troubleshoot database performance and issues.
Implement data quality standards and processes to ensure the accuracy and consistency of data.
Support data governance initiatives by ensuring compliance with data security, privacy, and regulatory requirements.
Continuously improve data management practices, ensuring the integrity and reliability of data.
Collaborate with data scientists, analysts, and other engineers to understand data needs and ensure data infrastructure supports their work.
Assist in the creation of dashboards, reports, and visualizations to support business decision-making.
Provide technical guidance and support to team members and other departments as needed.
Automate repetitive tasks related to data processing, extraction, and reporting.
Identify opportunities to improve processes and reduce manual interventions.
Document data pipelines, workflows, and processes clearly and comprehensively.
Provide reports on the status, progress, and performance of data projects.
Qualifications:
Bachelor’s degree in Computer Science, Engineering, Mathematics, or a related field (Master’s degree preferred).
Experience in data engineering, data management, or related roles.
Experience working with large-scale data processing systems and cloud technologies (AWS, Google Cloud, Azure).
Proficient in programming languages like Python, Java, Scala, or SQL.
Strong experience with data warehousing and ETL frameworks (Apache Spark, Hadoop, Airflow, Talend, etc.).
Hands-on experience with relational databases (e.g., PostgreSQL, MySQL) and NoSQL databases (e.g., MongoDB, Cassandra).
Experience with cloud platforms and big data services (AWS S3, Redshift, GCP BigQuery, Azure Data Lake).
Experience with data pipeline orchestration and automation tools (Apache Kafka, Airflow, Prefect).
Experience in data modeling, schema design, and data governance practices.
Strong analytical and problem-solving abilities.
Excellent communication skills, both verbal and written.
Ability to work independently and in a team-oriented environment.
Detail-oriented with a focus on quality and precision.
Ability to manage multiple tasks and meet deadlines.
Experience with machine learning pipelines and AI/ML model deployment.
Knowledge of containerization (Docker, Kubernetes).
Experience with data visualization tools like Tableau, Power BI, or Looker.
Knowledge of version control systems like Git.
Federal Experience is a plus.
Required Security clearance.
If you meet these qualifications, please submit your application via link provided in Linkedin.
Kindly do not call the general line to submit your application
Expected salary
Location
Albany, NY
Job date
Thu, 26 Dec 2024 23:45:46 GMT
To help us track our recruitment effort, please indicate in your email/cover letter where (jobs-near-me.eu) you saw this job posting.