segron logo

Is Manual Device Testing Hurting Customer Experience?

In 2012, an Open Signal study found that there were about 4,000 different Android device models on the market. Within a couple of years, that number had risen to 12,000, and it’s likely only gone up since then. As a device tester, you already know that there’s considerable diversity among your customers, and that their needs are going to vary on a case-by-case basis—but who knew there was so much diversity just in the devices themselves?

This degree of fragmentation has more than a few implications for testers. Most importantly, it means that ensuring a high quality of service for customers with a wide variety of different devices is a more Herculean task than ever.

Because there are so many mobile devices with which users could be trying to access your network and take calls, run apps, answer emails, etc., service verification is at once more important and more daunting than ever.

This begs the question: how can testers approach this changing landscape with an eye toward improving customer experience and ensuring a high quality of service? One potential answer is to ditch manual device testing.

Challenges in Mobile Device Testing

Like we alluded to above, the sheer degree of fragmentation within the mobile device market makes device testing a challenge. The 12,000 devices running Android’s OS only make up a fraction (albeit a large fraction) of a market that also includes iPhones, IoT (internet of things) devices, etc., which means that covering a majority of your users with your device testing involves running through thousands of different smartphones.

And performing functional tests for each one, potentially including 4G, 3G, 2G, and Wi-Fi verification. Considering that the average human tester can manually run through about 6-10 use case tests per day, this task can quickly seem impossible.

For many businesses, the solution of choice is to run simulated tests or perform tests on “rooted” or “jailbroken” devices, rather than utilizing out-of-the-box end user devices. Unfortunately, because these aren’t actually the devices that are being adopted by your customers, you open yourself up to potential gaps in verification.

If, for instance, you’re verifying a network update meant to reduce latency times for mobile broadband service, small differences in calibration will have a big impact on customer perceptions. As customer expectations with regard to things like latency and uptime become more and more stringent, the small differences between a simulated device and a real one will become magnified.

As we enter the 5G era, simulated tests will be even farther removed from reality, making it nearly impossible to verify the granular differences in network quality that many new technologies could come to depend on.

Customer Experience and Network Quality

Okay, so manual testing and simulated testing both present testers with real disadvantages when it comes to service verification in the modern marketplace. But how much should we really care? To what extent do your testing workflows directly impact your customers?

After all, if these issues didn’t correspond directly to customer experience, we might be able to skirt some of them a little more easily. Unfortunately, network quality and network verification tend to have a pretty direct impact on customer experience. Why?

Because device testing is the only way to uncover the bugs and service gaps that are viewed with increasing scorn by customers, whose expectations for network quality seem to grow every year. As those expectations increase, the pressure to test effectively increases along with it.

What this means, essentially, is that running-simulated tests, for instance, directly influences customer experience, often for the worse. By the same token, manual device testing—given the tremendous diversity of devices—is typically too time-consuming to hit desired test coverage levels, meaning that it, too, has the potential to negatively impact customer experience.

By committing to a testing framework that you know can’t achieve high levels of coverage, you are essentially committing to leaving many potential gaps in service quality unexamined, boosting the likelihood that your average user will encounter service failure or bugs trying to access your network with their device of choice.

Enter Device Testing Automation

This, perhaps unsurprisingly, is where automated testing comes in. Here, it’s crucial to draw a distinction between simulated tests and automated tests: simulated tests on virtual devices leave open the possibility that, when the rubber hits the road, your users will encounter issues that you weren’t able to replicate in your test lab; automated testing on out-of-the-box devices, on the other hand, alleviates this issue.

By reproducing something much closer to the actual conditions under which the devices will be used, you greatly increase the likelihood that you’ll find the bugs that are most relevant to your users. By the same token because automation can empower you to run hundreds of tests per day instead of only a handful, your testing coverage suddenly increases significantly.

Rather than focusing on only the most common devices, you can work to verify service on the latest technology as it emerges, potentially resulting in a competitive advantage in the eyes of early adopters.

In this way, automated testing begins to have a direct impact on customer experience. Rather than hoping that you’re able to cover the most common use cases and devices with manual testing, you’re capable of striving for much more complete coverage—leaving no stone unturned, so to speak.

Thus, the odds that a given customer will pick up the latest Samsung smartphone only to find that VoIP simply refuses to work with your service decreases considerably. This has a potential cascade effect, improving customer satisfaction and thus bolstering your customer retention rates. After all, the more effectively you can verify service across a wide variety of devices, the less likely users are to jump because of network issues.

Beyond that, automation makes continuous regression testing feasible (when time constraints might otherwise make it a hard thing to prioritize), meaning that as new devices enter the market you can continue to ensure that they fit into your service offerings. Trust us, your customers will thank you in the long (and short) run.


Interested in our Products ?

Scroll to Top
Segron logo black blue

Senior SaaS System Administrator

Technical Skills :
  • Oversee the sysadmin related tasks in our SaaS infrastructure (partially cloud based, partially bare metal)
  • Daily operation and maintenance of the system
  • Analysing and resolving incidents
  • Follow and help improving the incident and change management procedures
  • Design procedures for system troubleshooting and maintenance
  • Incorporating base OS updates and security patches
  • Ensure that systems are safe and secure against cybersecurity threats by raising change requests where potential threat is possible
  • Performing SW updates for the Segron SaaS SW stack (distributed architecture with clusters)
  • Configuring solutions like reverse proxy, firewalls, etc.
  • Building tools to automate procedures & reduce occurrences of errors and improve customer experience
  • Tutoring & coaching newcomers & less senior experts in the team
  • Interworking with the architects and IT admins of Segron to have the SaaS procedures inline with the Segron processes
Non-technical skills:
  • We are looking for a self-motivated, self-improving individual with a highly independent mindset and open and straightforward technical communication to help us to improve and maintain our cloud infrastructure of our powerful end-to-end testing solution ATF (Automated Testing Framework)
  • 3+ years hands-on experience with operation and monitoring of cloud / linux systems
  • 3+ years of hands-on experience with network devops elements: configuring routers, switches, networks
  • Hands-on experience with running live systems with infrastructure as a code mode of operation
  • Specific knowledge which brings direct advantage: Docker, Docker Compose, Grafana, Prometheus, Ansible, Debian Linux OS administration, Security
  • Experience in building and maintaining distributed systems (incl. redundancy, resiliency, load-balancing) is welcome
  • Excellent knowledge of English
Location :
  • Place of work: Bratislava (partially home office possible)
  • Rate: from 30 EUR/hour (possible higher rate, depends on experience)
Segron logo - The Next Generation of Active Testing
Segron logo black blue

CI/CD Senior Developer

Technical Skills :
  • A senior role with a proven expertise in software development, cloud computing, DevOps, and CI/CD
  • Experience in planning, designing, and overseeing the CI/CD strategy and architecture on the level of organization
  • Ability to tailor testing strategies which define and follow the best practices, standards, and policies for the software delivery process
  • Hands-on experience in creating and managing CI/CD pipelines and workflows (PaaC)
  • Ability to evaluate and recommend the best tools, technologies, and methodologies for the CI/CD implementation
  • Prior hands-on experience working with different CI/CD toolsets (Jenkins, Bitbucket, GitLab, artifactory, Ansible ..)
  • Proficient with DevOps tools API automation capabilities
  • Proficient with Atlassian Tools (BitBucket, Jira, Confluence) and agile SW development methodologies
  • Familiar with cloud patterns and best practices
  • Familiar with web performance best practices
  • Comfortable working in cloud DevOps ecosystem
  • Comfortable working with Linux platforms
  • Initial working experience in SW development is an advantage.
Non-technical skills:
  • Effective communication with technical as well and business stakeholders
  • Self-motivating, self-improving mindset
  • Ownership of relevant industry certificates is a plus
Location :
  • Location: Bratislava, Slovakia (with hybrid flexibility)
  • Rate: from 30 EUR/hour (possible higher rate, depends on experience)
Segron logo - The Next Generation of Active Testing
Segron logo black blue

Test Automation Engineer

Job description, responsibilities:

  • ATF system configuration, integration, operations & maintenance in customer environments.
  • Building tools to automate procedures & reduce occurrences of errors and improve customer experience.
  • Hardware Verification, Testing and Preparation within the Staging Process.
  • Contribution to customer and service partner technical support across multiple accounts by sufficiently managing priorities and deadlines for own work.
  • Segron Laboratory equipment configuration and maintenance support.
  • Hardware order and logistics support.
  • Problem analysis of ATF issues, troubleshooting and fault correction.
  • Interface towards SEGRON Development Team in case of product or software issues.
  • Interface towards the SEGRON Technical Sales Team to support planned activities.
  • System and Integration documentation and guidelines.
  • Perform root cause analysis for production errors.
  • Deployment of software updates and fixes.
  • Ability to work in a team environment serving multiple global customers.
  • Willing to travel for 3-5 days onsite deployments
Requirements/ Skills:
  • Excellent knowledge of English
  • Operating Systems: Linux, Windows, MacOS
  • Good Knowledge of Containers and Virtual Machines
  • Telco experience welcome
  • Python or other scripting experience or knowledge preferable
  • Educational Qualification: Computer Science/Engineering or work experience equivalent
  • Work Experience: 3-4 years preferred


  • Full time job (employment)
  • 3 days onsite, 2 days home office
  • Offered salary: from 1800 Euro (depends on seniority and skills level)
  • Variety of financial benefits
  • Place of work: Bratislava
Segron logo - The Next Generation of Active Testing
Segron logo black blue

Senior Python Developer

Technical Skills :

  • A solid, experienced SW developer with at least 10 years of experience in active SW development in different programming paradigms
  • Minimum 5 years of professional Python development experience
  • Master or college degree from Computer Science, Mathematics or STEM domain
  • Well educated in design and programming patterns that increase software’s efficiency and readability.
  • Very good analytical and problem solving skills.
  • At least three skills out of the following 4 skills are requested:
    • Microservices based architectures (Docker containers)
    • Linux
    • Ansible
    • Robot  Framework
  • Comfortable with sysadmin and DevOps skills (Ansible, YML files, Network Programming, IP protocols, designing and developing proxy servers for different protocols – example: streaming, integrating and compiling third party libraries on Linux (Debian))
  • Proficient with Atlassian Tools (BitBucket, Jira, Confluence) thorough understanding of Git and version control best practices
  • Familiar with cloud patterns and best practices
  • Familiar with web performance best practices

Non-technical skills:

  • Ability to work under pressure
  • Ability to abstract and explain your work
  • Strong understanding of Agile development process and experience working in an agile team
  • Strong communication skills with both technical and non-technical stakeholders
Location :
  • Bratislava, Slovakia (with hybrid flexibility)
  • Rate: from 35 EUR/hour (possible higher rate, but depends on experience)
Segron logo - The Next Generation of Active Testing