Category: Web Development

Category added for Website – 2017

Idyllic Named a Top Mobile App and Web Developer on Clutch!

Clutch Top App and Web Developers

Clutch an independent Research, Ratings and Reviews firm based in Washington D.C., has recently announced its list of top Web and App Development companies. The much coveted industry list is based on in-depth research, independent evaluation of projects executed and client interviews by Clutch analysts. Clutch regularly undertakes this meticulous exercise to help clients identify the best agencies and developers across the world.

Clutch ratings are based on genuine feedback from clients and the firms’ ability to deliver high quality scalable solutions. The diligent methodology covers various parameters and takes into account multiple factors to identify the best performing agencies. In recent times, the Clutch list has evolved into a reliable yardstick to evaluate a web and app development firm’s capabilities.

So it’s a matter of pride, as Idyllic is named as one of the best web development and app development firms on the list. Idyllic has consistently featured on Clutch’s Top Ruby on Rails Development, Web Development, App Development, Top Indian Firms lists.

At Idyllic Software, we are a leading design and development studio with a sharp focus on user research and design thinking, enabling us to move our clients’ businesses forward. Our clients are most important to us, and we work hard to build trusting, positive relationships with our clients. We are glad that this approach is paying off, as we were just named as one of the Top App Developers in Mumbai by Clutch in their annual research.

Clutch’s research covers mobile app development companies like us  around the world. Their complex research methodology evaluates companies on a number of qualitative and quantitative factors to determine the best firms. Most importantly, dedicated Clutch analysts speak directly with clients to gather information on how each company performs, producing full-length reviews for company profiles highlighting each company’s strengths and weaknesses.

Idyllic’s Clutch profile is full of positive feedback from our clients. As of now, we have received 11 positive reviews from our clients! Below are some excerpts from our happy clients:

“I was amazed at how quickly Idyllic could comprehend the business aspects of what we were building. They asked the right questions and didn’t waste our time. Idyllic gave us good inputs and helped us improve the design, rather than being a run of the mill company that builds software based on the requirements given. Idyllic never behaved like a contractor. They tried to solve the problem, working with us in order to achieve our goals and making sure that the business was satisfied.”

Director of Software Engineering, Sears

“Idyllic is a very specialized and talented Ruby on Rails shop, which was not a common model at the time. They have a solid understanding of the mechanics we use.”

CTO, SaaS Provider

“After spending time with Idyllic’s team and their founder, I’ve come to value their integrity and leadership. Whenever we had any concerns or questions, Idyllic’s Principal responded to us. I could tell that his team believed in his leadership and followed the integrity he set in terms of how to work with clients, how to follow their needs, what the code quality standard is, and so on. Idyllic’s Principal does an excellent job in exuding his passion for technology. I’ve come to look for the same elements in other relationships.”

Founder, Project Travel

We are proud that our clients have so much encouraging feedback to share on Clutch, and we look forward to growing our Clutch presence in the near future!



10 Essential Rules for WordPress Website Design

Wordpress Website Design

A great WordPress website does not just happen – it is a result of a lot of planning, strategy, and careful execution. It also requires you that you follow certain set principles that guide the entire process of designing and developing a website. Therefore, if you are looking to create a great WordPress website, here are some of the top 10 rules of WordPress Website Design that you must follow to get the best results:

  1. 1.Design For Your Audience

One of the most important principles of creating a WordPress website is that you should design the site with the users in mind. The user-centered design approach helps to give users an amazing user experience when they visit your site. Therefore, before you start the process of creating your WordPress website, you should first identify your audience and understand their needs.

  1. 2.Identify Your Brand

With so many businesses conducting their operations online, it is important that you develop a unique online brand identity that tells your website visitors who you are, what you do, what you are offering, where you are located, etc. Doing this helps to make it easy for them to recognize you and your business, thus increases the chances of them coming back again.

You can be able to create a brand identity by offering users your contact information – your address, phone number, email address, location, etc. You should also create a great “About US” section that features your brand story. On top of that, you should create a unique logo and develop a catchy memorable slogan. All of these will offer your brand a unique identity that users and potential customers can easily relate to.

  1. 3.Design For A Variety Of Devices

When users are accessing your site, they will do so with a range of devices – desktop PCs, laptops, Smartphone’s, tablets, etc. When designing your WordPress website, you should take this fact into account and design your site to offer a great user experience across all devices. It is also a good idea to design for multiple browsers, which will ensure that your site is accessible regardless of the browser being used.

  1. 4.Maintain Design Consistency

Design consistency is another top rule for WordPress website design that you should follow to the letter if you want a great website. With a consistent design, your visitors will enjoy a consistent user experience throughout the site. A consistent design also makes it very easy to navigate through all the web pages.

  1. 5.Great Use Of Colors And Images

In WordPress website design, the first impression is very important, as it can either capture the attention of your website visitors or cause them to turn back. Therefore, you need to make your design very appealing, which can be achieved by great use of color and images. You can also use other visual tools such as videos, animations, and infographics to make your website design more appealing.

  1. 6.Keep The Design Simple

Appeal and functionality are important when designing a WordPress website. However, this does not mean that you should cram your website with lots of features and design elements – doing this will only make the design appear chaotic, which is a huge turn off for the users. On top of that, using too many features will only make your design “heavy”, which will negatively affect the performance. Therefore, when you are building a website for your business, you should make it clear to your Web Development Company that you want a simple, clean design for your website.

  1. 7.Easy-To-Use Navigation Features

As a user, there is nothing more irritating than trying to navigate through a website without much luck. Therefore, when you are designing a website, navigation should be one of your top concerns, and you should ensure that users have an easy way of navigating through all the web pages.

  1. 8.Clear And Conspicuous Calls To Action

Calls to action are an important design element that can help to boost your website’s lead generation and conversion rates. As such, you should ensure that your website is designed with clear and persuasive calls to action that will persuade the user to take the intended action. On top of that, the calls to action should be designed with conspicuous colors and placed in a strategic location – easy to find and hard to miss.

  1. 9.Offer High-Quality Content

Content is one of the top design elements and it should never be overlooked when designing a WordPress website. With great content, you can be able to engage your website visitors with ease. Quality content will also help to improve the search engine rankings of your site, thus making your online brand more visible.  Therefore, while you are building your website, make sure that you also invest in top-notch content creation services that will provide you with quality, original content for your web pages.

  1. 10.Design With The Latest Trends And Practices

Another top principle of WordPress website design is to always stay current. This means that when you are building your website, you should make sure that only the latest technology, practices, and trends are used. By doing this, you will be able to ensure that your website is in peak condition. Using the latest design practices and trends will also help to make your website more competitive, which can be a huge boost for your online brand.

A Beginner’s Guide to Angular CLI

A Beginner's Guide to Angular CLI

What is Angular CLI?

For the uninitiated, Angular CLI is a Command Line Interface to build Angular apps quickly adhering to the best practices. Angular CLI has multiple advantages, some which are listed below for quick reference.
In this blog post, we’ll dwell on why we need Angular CLI, how to install it, how to use it i.e. the various commands that it offers to make our life easy. We’ll also look at how we can generate production ready builds using the angular CLI.

Why do we need Angular CLI?

  1. It creates and provides a scalable project structure
  2. It initializes a git repository and makes an initial commit.
  3. It handles all common tedious tasks out of the box.
  4. It generates initial HTML, typescript, CSS and unit test files.
  5. It creates a package.json with all the dependencies of Angular 2 and installs all the dependencies of node (npm install)
  6. It configures Karma to execute the unit tests with Jasmine
  7. It configures Protractor to execute the end-to-end tests.
  8. It creates all the files necessary to create a build of the application for production.

Let’s start with the basics

How to install Angular CLI ?

Please note that Angular CLI is a node package and it requires a node version > 4.

To check the node version on your machine. use the command – node -v

If you do not have node installed or you have node that is older than version 4. Then, you should consider updating the node. At the time of this blog post, the latest stable version of node is v7.9.0.

Once you have node installed, you also have the NPM (node package manager) installed. To check if you have npm installed, all you need to do is – npm -v

Now, we can install the Angular CLI using –

npm install -g angular-cli

Now we are ready to use Angular CLI

Using Angular CLI –

  1. ng new <<name_of_the_project>>

Creates an angular project with all the files, folders, packages, needed and npm install is executed.

Ex 1 – ng new hello-world (a project is created in the directory that you are currently on)

Ex 2 – ng new path/hello-world (Using a relative path)

Note –

Project name “01-test” is not valid. New project names must start with a letter, and must contain only alphanumeric characters or dashes. When adding a dash the segment after the dash must also start with a letter.

Project Structure – This is how an angular 4 project structure looks. 

  1. ng init

Creates an angular project in the current directory

mkdir name_of_the_project
// Creates a directory
cd name_of_the_project

ng init
  1. ng serve (also, ng server and ng s)

Serves an angular project via a development server. It’s very similar to “npm start”.

If the ng-serve is executed without any errors, then you should be able to check the result in localhost:4200/

Note – The default port used is 4200 and the live reload port used is 49512.

To use othert ports –

ng serve —port 4201 —live-reload-port 49153
  1. ng generate (also, ng g)

Creates Components, Services, Pipes, and other classes.

Scaffold Usage
Component ng g component my-new-component
Directive ng g directive my-new-directive
Pipe ng g pipe my-new-pipe
Service ng g service my-new-service
Class ng g class my-new-class
Guard ng g guard my-new-guard
Interface ng g interface my-new-interface
Enum ng g enum my-new-enum
Module ng g module my-module


  1. Using pre-processors using Angular CLI

Using style flag during the creation of the angular project

ng new my-project —style=sass
ng new my-project —style=less
ng new my-project —style=stylus

Using ng set – To add or modify the preprocessor in an existing project-

ng set defaults.styleExt scss
ng set defaults.styleExt less
ng set defaults.styleExt styl
  1. ng lint

To perform the js lint. To check for errors like – unused imports, unexpected spaces, extra comma etc.

If we declare the variable in the below manner and run ng lint –

myVariable : String; ❌

Expected no space before colon in property-declaration.

myVariable: String;

  1. ng test

Executes the jasmine unit tests in the spec.ts files and returns the details of the total tests run and those that failed.

  1. ng e2e

Runs the end to end tests written using the Protractor framework in the e2e folder

  1. ng build

Generates a build (development build, production build).

The build is generated inside the dist folder, which is ignored by the version control.

Creating a Development build

ng build --target=development --environment=dev


ng build --dev --e=dev


ng build --dev


ng build

Creating a Production build

ng build --target=production --environment=prod


ng build --prod --env=prod


ng build --prod

Note that the development build is neither obfuscated nor minified. Also, it doesn’t consider the browser caching which might lead to unexpected behaviour of the application.

  1. ng version

Returns the versions of angular-cli, node and the operating system.

  1. ng help

Outputs all the available angular cli commands.

All You Wanted to Know About Yarn Package Manager

Yarn Package Manager

Package management for the web has come a long way, there are many package managers like bower, npm etc. Of late, Yarn Package Manager has gained popularity going by the google search trends. 

What is Yarn and Why is it So Popular ?

Yarn is the new javascript package manager built by Facebook and Google together to solve some of the shortcomings of the popular framework NPM.

It doesn’t mean that Yarn is going to completely replace NPM but Yarn is a new CLI that fetches modules from npm registry. Nothing about the registry will change, you will still be able to fetch and publish packages as usual.

Before looking into the reasons why Yarn is getting so much attention, let’s have to look at the shortcomings of npm.

Problems with NPM:

I have listed below the drawbacks faced by Facebook and Google team. Chances are you’ve never encountered these problems with npm. We are going to compare npm and yarn, so you can decide which one is best for you.

  • Queued Installation: While getting dependencies from its repository, npm installs it one after another which takes up a lot of time.
  • Single Registry: NpmJS has lot of packages, but if the package you’re looking for is not in NpmJs, then you have to use a Cdn or other options
  • No Offline Installation: Npm doesn’t support offline installation as it doesn’t cache packages while installation. This means if you have installed an application before and try to install the same application offline, Npm gives an error

Yarn vs Npm – The Differences:

Yarn looks similar to Npm at first glance, but when you look closer and try using it, you will understand what makes yarn a great tool.

  1. Parallel Installation: While getting the packages from registry Npm and Yarn carry out a number of tasks. In Npm these tasks are done one after another. Yarn executes these tasks in parallel and reduces the installation time.
    Let’s compare the time Yarn and Npm take to install the same package. Before comparing you need to install Yarn globally, by using the below mentioned command.
 npm install -g yarn

Yes, I know it feels like using internet explorer to download Google Chrome 🙂
I will be installing express and gulp to compare the time difference between npm and yarn, you may try installing any package you like.

The Result:

  • Yarn: 32 seconds
  • Npm: 52 seconds

I know the difference for 2 packages is not much (20 sec) but imagine a project with a number of packages and the amount of time Yarn can save you. The bottomline, Yarn is faster.

  1. Yarn lock file: Imagine a scenario where you have created an app and it’s working as expected,  but after a while the app crashed due to package install version mismatch. To avoid this, yarn creates a lock file with all packages version when installed and updates the same on package changes.

In npm npm shrinkwrap command generates npm lock file and npm will read it before installation. The major difference here is that yarn creates and updates the lock file automatically but npm updates lock file only if it exists.

  1. Clear Output: If you run npm install command npm will list all installed packages, on the other hand yarn will display less commands like resolving packages, fetching packages, linking dependencies, building fresh dependencies with emojis (except for windows users) 🙁

Yarn Cli Commands:

There are two ways to install yarn, if you haven’t already done it. First is using the npm command below and second is from the official download page.

 npm install -g yarn

Note: Both methods need NodeJs already installed on your machine.

Yarn does not intend to replace npm as mentioned earlier. It uses the same package manager to install dependencies and save them to node_modules folder.

Initializing package.json using Yarn:

If you need new package.json file to initialize the dependencies tree you can use the command given below.

 yarn init

If you have used npm earlier, then this command will look familiar. This command will ask a similar question, that npm asks while initializing package files like  name, version, description etc.

Yarn Install:

In npm npm install command will install all the dependencies listed in the package.json file and also allow you to add other packages, However, with yarn there is a slight difference, yarn install command will only install packages from yarn lock file or package.json file. For adding or upgrading packages there is a different command in yarn.

Adding, Upgrading, and Removing Dependencies:

Adding: Similar to npm install [package name], the command yarn add [package name] allows you to add and install dependencies. As the name of the command implies, it adds dependencies, meaning it adds the reference of package in yarn lock and package.json file.

Yarn add command adds the dependencies to dependency section, if you need to add dependencies to the devdependency section you need to add --dev flag to the command.

Upgrading: This command yarn upgrade is similar to npm update and will update all the packages to their latest version. If you specify the package name and version number it will upgrade the package to that specific version.

For instance, yarn upgrade gulp@3.9.1 will update gulp to version 3.9.1.

Removing: This command will remove the package from node_modules folder and update the same in package.json and yarn lock file. yarn remove [package name].

For example, yarn remove gulp will uninstall gulp and update the same.

Stability and Future:

There were some issues at the time of Yarn launch but the rate of resolved issues is astounding. It indicates that the community is working hard to find and resolve the bugs. Yarn appears stable for most users, including yours truly. I have been using yarn for more than two and a half months and haven’t faced any issue.

The future looks bright, as it is backed by Google and Facebook. Yarn will get actively developed which will make it the default package manager or get forked and used as the official npm package manager.

But nonetheless, the project looks very promising and I would definitely recommend trying Yarn on single projects.


In the final analysis, Yarn scores way higher than npm. We are getting fast package installation, lock file and much more. You can try it on a pilot project and see if it works for you. Atleast give it a shot, it’s worth a try.

And for those of you, using yarn already in your projects please share your experience with Yarn so far. Let me know your take in the comments below.

Idyllic Featured as one of the Top Ruby on Rails Developers

Top Web Developers

Clutch, an independent ratings and reviews firm based in Washington, D.C., has featured Idyllic as a leader in Ruby on Rails web developers, giving Idyllic the #2 rank in Rails technology.

Clutch evaluates hundreds of web developers across the globe. Their methodology is based on a number of qualitative and quantitative criteria, including previous work, market presence, client accolades, among others.

Most importantly, the backbone of Clutch’s scoring methodology comes from verified client interviews of our web development services, in which Clutch speaks directly to our clients to get a sense of the services provided.

Here are some client quotes in the interviews conducted by Clutch:

One client, the founder of a travel abroad SaaS company, stated:

“They delivered good results and didn’t simply code based on their individual egos, giving us things which we didn’t need. Idyllic worked based on our business goals, listened to our instructions, and took initiative appropriately.”

When asked about his overall satisfaction with our web development services, he also added:

“After spending time with Idyllic’s team and their founder, I’ve come to value their integrity and leadership. Whenever we had any concerns or questions, Idyllic’s principal responded to us. I could tell that his team believed in his leadership and followed the integrity he set in terms of how to work with clients, how to follow their needs, what the code quality standard is, and so on. Idyllic’s principal does an excellent job in exuding his passion for technology. I’ve come to look for the same elements in other relationships.”

Another client said:

“The marketing team was thrilled at how rapidly improvements and changes were being made to the website. Our internal engineers also liked the code quality delivered by Idyllic. I considered them a lifesaver, especially since I was responsible for other projects while keeping internal customers happy.”

As 2016 came to a close, Clutch also recognized Idyllic as one of the Top Web Developers in India in their research.

We’re excited to be included in Clutch’s research, which highlights our focus on providing high-quality work with great customer focus and ranks as a top web developer. A shout-out to all our enterprising clients who expressed their faith in us and Clutch for their in-depth interviews and diligence. For more information, do check out Idyllic’s reviews on Clutch.

6 Best Practices for Leading a Development Team

When you are the team lead of a small or a big development team, you have many responsibilities on the project that you don’t as a developer. You become a key player in ensuring successful delivery of the project. Here are 6 things to keep in mind as a lead.


This is one of the key aspects of your responsibility as a team lead, Visualization of the application architecture. This will help you every time you want to extend the application or accommodate  any changes. It will also help you guide the development team in the part of the application (s)he is working on.

On top of everything when the development is in full swing, you will be able to refer your architecture diagram to ensure you are adhering to the software engineering principles you had in mind while creating the architecture.

Code Quality

This is another very vital responsibility that one needs to execute. You will be making sure the overall quality of the codebase is maintained. This can be achieved by doing a few things.

  1. Code Reviews
  2. Defining Coding Standards
  3. Knowledge sharing in the team

Making Progress

As a developer, you are usually tempted to revisit your code and come up with a better design or maybe you want to refactor the code you have written. At times the development team may go overboard with this refactoring and over engineer a solution. As a lead, you have to keep track of the deliverables and the cost the business is incurring. At times a good solution is feasible and, you need to draw the line for the development team keeping the business prospects in mind.

Planning and Process

Being a lead you will be one of the custodians of the process. You will need to ensure that the team is following the process strictly. This will help you stay updated with the current state of the project. At the same time, you can fix ownership and accountability of the project. With a solid execution plan, you can manage expectations from the business well. The lead should spend some time to periodically revisit the roadmap.


As a lead you are involved in two-sided communication. You are presenting the business side of story to the development. At the same time you are presenting the development team’s perspective and challenges to the business. If you can communicate effectively on both the fronts you will be able to ensure smooth execution of the project.

Writing Code

As a lead, you are always in a dilemma whether you should be writing code or you should be mentoring the team and be involved in the planning. The simple answer is YES you should be writing code (a certain Zuckerberg still does). The advantage is that you are always up-to-date with the codebase. You avoid the situation of making uninformed decisions. Also, as a good leader, you should work with the team in trenches. At heart you are still a developer, you will enjoy this part the most and it will refresh your mind.


Build your own mini Heroku powered by Docker in 3 simple steps

Heroku, as you are most probably aware of, is a popular PaaS (Platform as a Service) that allows developers to easily deploy their applications developed using different languages or frameworks to the cloud using a simple ‘git push’. This saves you from having to provision a server yourself for deploying your app which involves installing all the required dependencies on the server, checking config files etc. There are tools to automate server provisioning and deployment for example Vagrant, Chef, Puppet, Ansible, F*cking Shell Script etc, but they have their own learning curve. A PaaS like Heroku takes care of all the deployment for you, literally taking care of the ops in devops so that you can focus on developing great apps rather than on also building the required plumbing. However, this power comes at a cost. They do have cheaper plans for hobbyists but deploying any serious application using Heroku today costs a minimum of $50.


However, if you don’t need all the extras that Heroku offers and just need the simple git push based deployment and Buildpack support then there is an open source alternative called Dokku that anyone can run on their own Linux server running the 64 bit version of Ubuntu 14.04, and have their own PaaS up and running in minutes. For example, if you are using DigitalOcean, you can have your own PaaS up and running in minutes for as low as $5 a month. For the curious, Buildpacks are what Heroku uses to support the automatic deployment of applications developed using different languages or frameworks like Rails, Node.js, Java, PHP etc and there is a separate build pack for each platform. Since all the official build packs have been open sourced by Heroku, Dokku is able to use all of them and you can view the list of those here.


Dokku describes itself as ‘Docker powered mini-Heroku’ and is written in around 200 lines of bash script !. It uses docker to emulate the Heroku-like experience. And though you don’t really need to know a lot about docker and containers to get started with Dokku, however being familiar with them will help you understand how Dokku works and more effectively use it for your own purposes. So before I demonstrate Dokku I will provide a brief introduction.


Docker is the software that allows you to create and manage containers. Containers are analogous to virtual machines in the sense that they allow you to isolate multiple processes or applications. But unlike virtual machines which each have their own copy of the entire operating system, containers share the operating system kernel of the host they are running on, which allows multiple containers to run at the same time on the host operating system which makes them extremely fast and resource efficient as compared to virtual machines.This is made possible by Linux kernel features like LXC and layered file systems like AuFS.  Also since each application is packaged with the entire runtime and dependencies it needs to run, in a container, it can be deployed as is on any system with Docker installed so it doesn’t matter which system you used to develop your app and which one you are deploying it on making your apps truly portable as in ‘write once run anywhere !’ .


Now container based virtualization support has been available in Linux kernel since long but it was not simple to use.Docker has made containers easily accessible to everyone.However, Docker alone does not allow for the easy Heroku-like deployment. And this where platforms like Dokku come in which stand on the shoulders of giants and use Docker under the hood and also various other open source projects like gitreceive, build step and Nginx to provide the PaaS experience.


To demonstrate Dokku I will deploy a simple Node.js application to a Linux VPS running Ubuntu 14.04 x64 in 3 steps.


Step 1. On your local machine run the following command to fetch the sample node.js app.
Step 2.  (One time step) On your server install dokku using the following two commands as a user with Sudo access. If you are using Digital Ocean then you can entirely skip these two commands and use their one-click install to create a server with Dokku pre-installed.
Then just navigate to your server ip and add your domain (if you have one) and add your public ssh key to finish the configuration.


Step 3. And finally, the below 2 commands in your application directory to deploy the app !.You can change my-node-app to whatever another name you want.


And voila, within a few minutes your app would be up and running at ‘http://your-domain-or-server-ip-here’. It’s really that simple!.


To get a peek behind the scenes and confirm that the application is really being run inside a container, run the command ‘docker ps‘ on the server and see the container running the nodejs application.


For more information on dokku checkout Also, there is another project called dokku-alt, stands for dokku-alternative, which is a fork of dokku and comes with a lot of commonly used plugins like data stores etc preinstalled. Hope this post has whetted your appetite for Dooku in particular and also docker and container technology in general and you would explore it in more detail. If you need further help or clarification on anything above then you can reach out to me at Happy Hacking !.

Working with Logstash

Logstash a centralized tool to collect and aggregate logs. It is so intuitive and it’s configuration are so easy to understand that you would just love it.

The post describes how to work with Logstash and Logstash configuration.

In nut shells, Logstash is composed of three main components.


Input :  What is the medium/source through which Logstash would receive your log events.
A valid input source could be stdin,tcpudpzeromq etc. In fact, Logstash has a wide range of input tools which you can choose from.(to get full list input plugin click here)

The input block essentially looks like this.

input {
   stdin {
      codec => 'plain'


Output : The source or medium to which the Logstash would send or store it’s event.
Just like input Logstash provide a wide range of Output plugin as well.

The vanilla output block looks like this

output {
   stdout {
      codec => 'rubydebug'

If you really aren’t considering to perform any filtration on data or log message you receive, most of the times the above blocks(input and output) is sufficient to start with Logstash.

Note: We are making a minor adjustment in our working example. Instead of using the stdin we would be using tcp as the input plugin.

A final look at our configuration.

## logstash.conf
 input {
   tcp {
      port => '5300'

output {
   stdout {
      codec => 'rubydebug'


Testing Configuration –

logstash -f logstash.conf --configtest


Loading the Logstash

logstash -f logstash.conf


You might get a little help from the below screenshots to understand how Logstash output looks like.

Note: I had used Telnet to send logs to Logstash.

@timestamp: An ISO 8601 timestamp.
message: The event's message. 
@version: the version of the event format. The current version is 1.
host: host from which the message / event's was sent. 
port: port of the client.


Filter Filter plugin, are used to massage(filter) the logs(if needed) so that one modify the received log message before output(ting) it via output plugin.

Simple a filter block look like this. (we will explore this in our next example)

filter {
   grok {
     ## grok filter plugin 


To explain the power of Logstash, let us just work with a demo example.

Here we have an application which generates logs of various types

– Custom debugging logs.
–  SQL logs etc.


[20-JUN-2016 14:00:23 UTC] Received Message

[20-JUN-2016 14:00:24 UTC] Before query the IP Address
(1.0ms)  SELECT "ip_addresses"."address" FROM "ip_addresses" WHERE "ip_addresses"."resporg_accnt_id" = 3
[20-JUN-2016 14:00:24 UTC] After query the IP Address
[20-JUN-2016 14:00:24 UTC] The Ip address found is X.X.X.X

[20-JUN-2016 14:00:27 UTC] Quering ResporgID
ResporgAccountId Load (2.0ms)  SELECT resporg_account_ids.*, tfxc_fees.fee as fee FROM "resporg_account_ids" AS resporg_account_ids LEFT JOIN ip_addresses ON = ip_addresses.resporg_accnt_id LEFT JOIN tfxc_fees ON = tfxc_fees.resporg_account_id_id WHERE "resporg_account_ids"."active" = 't' AND (((ip_addresses.address = 'x.x.x.x' AND ip_addresses.reserve = 't') AND ('x.x.x.x' = ANY (origin_sip_trunk_ip))) OR (resporg_account_ids.resporg_account_id = 'XXXX') OR (resporg_account_ids.resporg_account_id = 'XXXX'))
[20-JUN-2016 14:00:27] Resporg ID is TIN

[20-JUN-2016 14:00:29 UTC] Querying Freeswitchinstance 
FreeswitchInstance Load (1.0ms)  SELECT  "freeswitch_instances".* FROM "freeswitch_instances" WHERE "freeswitch_instances"."state" = 'active'  ORDER BY "freeswitch_instances"."calls_count" ASC, "freeswitch_instances"."average_system_load" ASC LIMIT 1
[20-JUN-2016 14:00:29 UTC] FreeswitchInstance is IronMan.

[20-JUN-2016 14:00:29 UTC] Get the individual rate
IndividualCeilingRate Load (0.0ms)  SELECT  "individual_ceiling_rates".* FROM "individual_ceiling_rates" WHERE "individual_ceiling_rates"."resporg_account_id_id" = 7 AND "individual_ceiling_rates"."originating_resporg_id" = 3 LIMIT 1
[20-JUN-2016 14:00:29 UTC] The individual rate is 20

[20-JUN-2016 14:00:30 UTC] Query the individual rate
Rate Load (1.0ms)  SELECT  "rates".* FROM "rates" WHERE "rates"."resporg_account_id_id" = 3 LIMIT 1
[20-JUN-2016 14:00:30 UTC] The Selected rate is 40


Now, we need our system to output(or store) the logs based on their type(SQL and Custom type)

This is where the power the Filter(plugin) outshine.

GROK filter plugin

A closer look at filter(grok) plugin suggests that one can add a regex for the incoming log events(for filtering).
Note: Grok has a wide range of regex pattern (120+) that you can choose from. But it’s power is not limited to predefined regex pattern. In fact, one can provide a custom regex pattern as well (like in our case)
In our cases, we can apply regex on either SQL or Custom logs(we are choosing SQL message) and then segregate them.
Note. If you need help building patterns to match your logs, you will find the grokdebug and grokconstructor application quite useful.

The Regex –

Let’s define our configuration now.

## input the log event via TCP.
input {
   tcp {
      port => '5300'

filter {
  ## apply this filter only to log event of type custom
  if ([type] == "custom") {
    grok {
       ## load your custom regex pattern 
       patterns_dir => "./pattern"
       ## Compare the message with the you applied regex
       match => { "message" => "%{ARSQL:sql}" }
       ## if the message matched the given regex apply a field called "grok" match
       add_field => {"grok" => "match"} 

  ## if the field has a grok match, which means that  above regex match
   if ([grok] == 'match') {
      ## apply mutate filter plugin to replace the type from CUSTOM to SQL
      mutate {
        replace => {"type" => "sql"}
        ##  remove the grok field that was added in the earlier filter
        remove_field => ["grok"]

## output plugin. For now we will be using rubydebug but we can every easily used any of the output plugin 
output {
   stdout {
      codec => 'rubydebug'


Let examine output

       "message" => "Received Message",
      "@version" => "1",
    "@timestamp" => "2016-06-20T14:00:23.320Z",
          "host" => "werain",
          "type" => "custom" ## custom tag

       "message" => "(1.0ms)  SELECT "ip_addresses"."address" FROM "ip_addresses" WHERE "ip_addresses"."resporg_accnt_id" = 3
      "@version" => "1",
    "@timestamp" => "2016-06-20T14:00:24.520Z",
          "host" => "werain",
          "type" => "sql" ## we have successfully managed to change the type to sql(from custom) based 
                          ## on the grok regex filteration

Notice the type SQL being mutated(replaced) in place of a custom type.
Note:  Well if that is not enough you can ask a LogStash to filter the event from an external program.If you want you simply try my demo example and LogStash configuration defined over here and here

That all folks. I hope I manage to do justice to the amazing library called LogStash which has simplified my tasks of log-management to such ease.



Use of Amazon SES in Rails App to Send emails

What is Amazon SES?

Amazon SES is an email platform that provides an easy, cost-effective way to send and receive email using your own email addresses and domains.

Sending Email with Amazon SES using SMTP

You can use Simple Mail Transfer Protocol(SMTP) to send emails by using Amazon SES. You need an Amazon SES SMTP username and password to access the Amazon SES SMTP interface.

To create SMTP credentials:

1. Sign into the AWS Management Console and open the Amazon SES console at

2. In the navigation pane, click SMTP Settings.

3. In the content pane, click Create My SMTP Credentials.


4. In the Create User for SMTP dialog box, you will see that an SMTP username has been filled in for you. You can accept this suggested username or enter a different one. To proceed, click Create


5. Click on Show User SMTP Credentials. Your SMTP credentials will be displayed on the screen; copy them and store them in a safe place. You can also click Download Credentials to download a file that contains your credentials.

After creating your SMTP credentials, open the Verified Senders/Email Addresses screen. Before you can send an email using Amazon SES, you have to own an address or addresses which are going to be used as senders across the SES SMTP mail servers.


Configure mailer with SMTP with rails app:

Add the following code in config/environments/*.rb

Then add following code in mailers/ses_mailer.rb

Note: If your account is in Amazon SES sandbox, then you must also verify the email address of every recipient. For moving out of the sandbox, see Moving Out of the Amazon SES Sandbox.

View: view/ses_mailer/welcome.html.erb

Call Mailer:

You can track your email sending statistics on Sending Statistics of Amazon Console.

Sending emails using aws-ses gem:

1. Add aws-ses to the gem file –

2. Extend ActionMailer in config/initializers/your_config_file.rb, where config_file.rb is the name of the file which contains initialization routines for ses gem:

3. Then set the delivery method in `config/environments/*.rb` as appropriate:

You are now ready to send emails from rails app through Amazon SES. Just make sure you have verified email addresses through which you are going to send the emails and you are out of Amazon SES sandbox. Happy Mailing!!

WebRTC Opentok in Ruby On Rails for Text chat

This post will take you through adding text chat functionality using Opentok (a WebRTC platform) in Ruby on Rails. OpenTok is a WebRTC platform for embedding live video, voice, and messaging into your websites and mobile apps. Opentok has its own signalling server. We will be using the OpenTok signaling API.
First, we will make the basic setup for using Opentok service. Here we will be using ‘opentok’ ruby gem.

1) Add opentok gem to your Gemfile.

gem ‘opentok’


2) Then create a session that will attempt to transmit streams directly between clients.

Store session id and token somewhere in the database, so that you can use it later on.
Here token gets expired in 30 days(default value). You will need to regenerate after 30 days.

3) Now it’s time to connect to session that we have created above, Add this to your HTML page,

A signal is sent using the signal() method of the Session object. One can receive a signal by listening to a signal event dispatched by session object.
So, All clients who are connected to the same session and listening to same signal type(‘text_chat’ here) will receive a message as soon as someone publishes to channel ‘text_chat’.

Subscribe To Our Blog

Get access to proven marketing ideas, latest trends and best practices.

Next up home


Lets build cool stuff

Share your contact information & we will get in touch!

I want (Tell us more about your dream project)