Home » Antivirus » Securing web sites from web hackers

If you sell some stuff, it makes sense that headlines of what you are selling spreads far and wide to invite as many customers as possible. To this end, products selling websites want to be found near the top of popular search engine and to be comprised on price comparison websites. It is recognized that computerized software robots (bots) must access web sites to attain this, counting the web crawlers used by search engines and web scrapers used by price appraisal sites; these are supposed good-bots.


But, not all bots are good and, as the digital platform provider Datalex found, some bots can be very bad really. The company offers a unified e-commerce platform for combining pricing, travel operators, shopping, analysis for journey bookings and order management, all this across the number of access channels would be travellers wish to use. Its major customers include Brussels Airlines, Virgin Atlantic, Swiss International Air Lines and Lingus. It has many more customers across the globe.

Datalex allows its clients, which are majorly travel operators, to manage difficult personalized bookings for customers. As well as the actual ticket for a journey this might comprise increased baggage allowances, lounge access, seat upgrades, in-flight meals, flights, car hire, hotels, travel Insurance, field transportation and so on.

The pain is that such info is not just of interest to genuine travellers planning their journeys and gentle good-bots. Dishonest competitors use web hackers to steal information from travel sites and use it on their own websites that can adversely influence search engine optimization, and to control and undercut prices. To deal with the hackers, simply install a Norton Antivirus. To get any assistance regarding antivirus, simply call toll free number of Norton Antivirus tech Support Company.

Web hacking activity can be tenacious and hit the performance of the sub-domains and drive up back end costs as responsibilities are run up for call-outs to other facilities, which are created both by genuine users and bad bots. Combined across the Datalex platform this can develop a difficulty for all the clients it hosts, even the ones that are not being straightly targeted.

Justifying web hackers is complicated as you do not want to chunk the good ones. In a recent report, Quocirca looked at the issue of distinguishing good-bots from bad-bots and managing their activity.

There is a procedure called the robot exclusion standard/protocol that is used by good bots to check which parts of a website they are welcome to surf; but, this relies on procedure and bad-bots will just overlook it. Blocking the IP addresses manually of bad-bots is annoying as it is simple for the criminals to just move their bad bots to new locations. As most bad-bots copy genuine user behaviors and it is difficult for web application firewalls that focus on irregularities and susceptibilities, to discover them. Login enforcement, “are you a human?” and strong verification tests are all interruptions for genuine users and good-bots.

The answer for Datalex in the end was professional bot discovery and mitigation technology from a seller called Distil Networks. The goal is to use an opposite web proxy to detect bots straight through a range of methods including behavioral analysis, machine learning and digital fingerprinting. Bots can then be separated and policies applied; good bots can be white-listed and bad-bots, as well as unwanted hackers, blocked. It removes the annoying hits against its clients’ sites, making them steadier and reducing backend infrastructure costs. On average, removing bad bots reduced traffic to Datalex customer sites by twenty percent with no impact on real human users.