Data quality matters when building and refining a Classification Model

In the world of machine learning, data takes centre stage. It’s often said that data is the key to success. In this blog post, we emphasise the significance of data, especially when building a comment classification model. We will delve into how data quality, quantity, and biases significantly influence machine learning model performance. Additionally, we’ll explore techniques like undersampling as […]

Use pytest to run Great Expectations checkpoints

At Qxf2, we’ve successfully integrated Great Expectations into majority of our projects. We now have GitHub workflows in place to run Great Expectations checkpoints before deploying our applications to production. However, as our test suite expanded, we encountered a few challenges: 1. Triggering valid checkpoints. 2. Aggregating checkpoint results. To address these issues, we turned to pytest. In this post, […]

Investigation of the application deployed on Kubernetes

As a tester, we work with applications deployed on Kubernetes. That means, we need to know how to interact with various components of Kubernetes. But most online tutorials start with stuff that applies mostly to developers and DevOps engineers like install, writing deploy scripts, etc. Those are not really useful to testers, at least not directly. So, in this blog, […]

Build a semantic search tool using FAISS

This post provides an overview of implementing semantic search. Why? Because too often, we notice testers skip testing more complex features like autocomplete. This might be ok in most applications. But in domain specific applications, testing autocomplete capabilities of the product is important. Since testers can benefit from understanding implementation details, in this post, we will look at how autocomplete […]

Getting everyone to write infrastructure tests

In this post, we will outline our usual strategy to get an entire team to quickly fill testing holes – like writing infrastructure tests. Why? Because we find ourselves using this approach frequently at clients. Some context: Qxf2 engineers work with early stage products that neglected writing tests in favour of shipping quickly. This means, we almost always enter situations […]

Integrating CloudWatch Alarms with Skype

At Qxf2, our AWS environment hosts a multitude of applications, so monitoring the AWS services in real time is crucial for maintaining system reliability and performance. We’ve been using CloudWatch alarms to help us watch over our resources, these alarms used to send us emails whenever something went wrong. The problem was, it is often easy to miss these alerts […]

Auditing AWS cloud resources with Chef InSpec

Continuing from the preceding blog, ‘Auditing OS Level Resources with Chef InSpec’ which delved into utilizing the Chef InSpec open-source tool for testing individual servers via OS level resources. Now, we embark into the domain of Chef InSpec’s facility in cloud environments. Chef InSpec extends its support to cloud platforms like AWS, Azure, and GCP. Referring to insights shared in […]

Exploratory Testing with Logmine

This post will discuss how to improve your Exploratory Testing using Logmine. For sometime, I have been looking out for a log analyzer when I started doing exploratory testing with a new product. Logs are a gold mine of information if used well. You can infer common problems, get testing ideas, understand typical behaviour of a product and more from […]