Data and Integration Engineer - I

Attain

  • Bangalore, Karnataka
  • Permanent
  • Full-time
  • 6 hours ago
  • Apply easily
Company DescriptionFounded in 2018 in Bangalore, the center of India's high-tech industry area, Attain has grown to serve a global client base of startups, early stage companies, and SMEs.Attain is US based software company TANGO's India partner. TANGO's software solution helps customers find, build and manage retail and office locations in a single suite/ software platform. Attain supports Tango with services in the areas of development, implementation, QA, and client support.Job Description
  • Assist in configuring and maintaining integration workflows using middleware tools to support data exchange between Tango's API architecture and third-party systems.
  • Support ETL processes, including data extraction, transformation, and loading, following Tango's proprietary data migration methodologies—incorporating data analysis of input, internal, and output message data.
  • Participate in testing and validating integrations, ensuring data quality, asynchronous processing, and polling-based synchronization meet client requirements; plan and lead integration tests with external systems.
  • Collaborate on low-touch implementations by leveraging standard API endpoints and flat file transfers (e.g., SFTP-based), deploying standard integrations, and providing daily operational support.
  • Provide Level 3 Production Support for integration-related issues, including root-cause analysis and remediation within defined SLAs; serve as the primary coordinator for internal engineering resources.
  • Contribute to documentation updates for integration playbooks, Swagger files, user guides, test procedures, performance specifications, and product manuals to enhance self-service capabilities.
  • Assist in estimating Level of Effort (LOE) for custom integrations during SOW/CO development and client engagements; prepare data templates based on Source to Target Mapping (STM) documents.
  • Perform data transformations, including merging, ordering, aggregation, and resolving data cleansing/quality issues using various data processing tools; add/remove columns in ETL scripts (e.g., Kettle) using STM/Data Templates.
  • Run lookups queries, update lookups, execute data quality scripts, format and validate data quality reports, and run scripts on data in staging databases while updating data load checklists.
  • Conduct internal smoke testing of loaded data in the Tango application and prepare/add/remove columns in sample Business Data validation trackers using STM.
  • Integrate information from multiple data sources, solving common transformation problems, and effectively communicate with managers and project teams regarding data sets and reporting needs.
  • Engage in agile iterations for refining transformation routines and business rules, prioritizing critical path data elements while understanding business-domain context to clean and transform data for analysis.
Qualifications
  • Bachelor's degree in Computer Science, Information Technology, Data Science, Mathematics, Statistics, or a related quantitative field (or equivalent experience).
  • 0-2 years of professional experience in software development, integrations, or data wrangling (internships or academic projects count); at least 2+ years of SQL proficiency.
  • Proficiency in JavaScript, with a solid understanding of scripting for data manipulation and automation; 2+ years of experience with Kettle (Pentaho) or similar ETL/data processing tools.
  • Hands-on experience in any ETL/Reporting tool and manual data analysis using Excel or other spreadsheet tools.
  • Basic knowledge of RESTful APIs, asynchronous processing, and data formats (e.g., JSON, CSV); understanding of both waterfall and agile project management methodologies.
  • Proficient in Microsoft Office software applications, such as Word, Excel, and PowerPoint.
  • Excellent analytical and problem-solving skills with the ability to perform root-cause analysis; strong attention to detail and prioritization skills to handle multiple tasks in a fast-paced environment.
  • Strong written and verbal communication skills; aptitude to learn new technologies.
Preferred Qualifications:
  • Experience with Node.js for building scalable backend services and handling data pipelines; 3 years of scripting experience for data analysis in languages like SQL, Python, or R.
  • Familiarity with Node-RED or similar low-code/no-code flow-based tools for middleware and integration workflows.
  • Demonstrated experience in data wrangling, including extracting and transforming different types of data files (e.g., CSV, XLSX) into analyzable structures using reproducible coding and scripting techniques.
  • Exposure to ETL tools, data migration processes, or middleware architectures (e.g., Node-RED in a production environment).
  • Understanding of security protocols such as JWT authentication, PGP encryption, or SAML 2.0 for SSO.
  • Prior experience in SaaS environments, particularly IWMS or enterprise software integrations.
Additional Information
  • Need for immediate joining.
  • Work from home.

Attain

Similar Jobs

  • Senior Software Engineer

    Ferguson

    • Bangalore, Karnataka
    About Ferguson Ferguson is the largest value-added distributor serving the specialized professional in the residential and non-residential North American construction market. We …
    • 21 hours ago
    • Apply easily
  • Data Tagging QA Specialist

    YipitData

    • Bangalore, Karnataka
    Job Description About Us: YipitData is the market-leading data and analytics firm. We analyze billions of data points every day to provide accurate, detailed insights across in…
    • 22 hours ago
    • Apply easily
  • Senior Data Analyst

    Delta Air Lines

    • Bangalore, Karnataka
    About Delta Air Lines About the Company Delta Air Lines (NYSE: DAL) is the U.S. global airline leader in safety, innovation, reliability and customer experience. Powered by our…
    • 22 hours ago
    • Apply easily