Network


Latest external collaboration on country level. Dive into details by clicking on the dots.

Hotspot


Dive into the research topics where Pawan Prakash is active.

Publication


Featured researches published by Pawan Prakash.


international conference on computer communications | 2010

PhishNet: Predictive Blacklisting to Detect Phishing Attacks

Pawan Prakash; Manish Kumar; Ramana Rao Kompella; Minaxi Gupta

Phishing has been easy and effective way for trickery and deception on the Internet. While solutions such as URL blacklisting have been effective to some degree, their reliance on exact match with the blacklisted entries makes it easy for attackers to evade. We start with the observation that attackers often employ simple modifications (e.g., changing top level domain) to URLs. Our system, PhishNet, exploits this observation using two components. In the first component, we propose five heuristics to enumerate simple combinations of known phishing sites to discover new phishing URLs. The second component consists of an approximate matching algorithm that dissects a URL into multiple components that are matched individually against entries in the blacklist. In our evaluation with real-time blacklist feeds, we discovered around 18,000 new phishing URLs from a set of 6,000 new blacklist entries. We also show that our approximate matching algorithm leads to very few false positives (3%) and negatives (5%).


international conference on computer communications | 2013

On the impact of packet spraying in data center networks

Advait Dixit; Pawan Prakash; Y. Charlie Hu; Ramana Rao Kompella

Modern data center networks are commonly organized in multi-rooted tree topologies. They typically rely on equal-cost multipath to split flows across multiple paths, which can lead to significant load imbalance. Splitting individual flows can provide better load balance, but is not preferred because of potential packet reordering that conventional wisdom suggests may negatively interact with TCP congestion control. In this paper, we revisit this “myth” in the context of data center networks which have regular topologies such as multi-rooted trees. We argue that due to symmetry, the multiple equal-cost paths between two hosts are composed of links that exhibit similar queuing properties. As a result, TCP is able to tolerate the induced packet reordering and maintain a single estimate of RTT. We validate the efficacy of random packet spraying (RPS) using a data center testbed comprising real hardware switches. We also reveal the adverse impact on the performance of RPS when the symmetry is disturbed (e.g., during link failures) and suggest solutions to mitigate this effect.


acm special interest group on data communication | 2011

On the efficacy of fine-grained traffic splitting protocolsin data center networks

Advait Dixit; Pawan Prakash; Ramana Rao Kompella

Multi-rooted tree topologies are commonly used to construct high-bandwidth data center network fabrics. In these networks, switches typically rely on equal-cost multipath (ECMP) routing techniques to split traffic across multiple paths, such that packets within a flow traverse the same end-to-end path. Unfortunately, since ECMP splits traffic based on flow-granularity, it can cause load imbalance across paths resulting in poor utilization of network resources. More fine-grained traffic splitting techniques are typically not preferred because they can cause packet reordering that can, according to conventional wisdom, lead to severe TCP throughput degradation. In this work, we revisit this fact in the context of regular data center topologies such as fat-tree architectures. We argue that packet-level traffic splitting, where packets of a flow are sprayed through all available paths, would lead to a better load-balanced network, which in turn leads to significantly more balanced queues and much higher throughput compared to ECMP.


symposium on reliable distributed systems | 2010

Shedding Light on Enterprise Network Failures Using Spotlight

Dipu John; Pawan Prakash; Ramana Rao Kompella; Ranveer Chandra

Fault localization in enterprise networks is extremely challenging. A recent approach called Sherlock makes some headway into this problem by using an inference algorithm over a multi-tier probabilistic dependency graph that relates fault symptoms with possible root causes (e.g., routers, servers). A key limitation of Sherlock is its scalability because of the use of complicated inference algorithms based on Bayesian networks. We present a fault localization system called Spotlight that essentially uses two basic ideas. First, it compresses a multi-tier dependency graph into a bipartite graph with direct probabilistic edges between root causes and symptoms. Second, it runs a novel weighted greedy minimum set cover algorithm to provide fast inference. Through extensive simulations with real service dependency graphs and enterprise network topologies reported previously in literature, we show that Spotlight is about 100× faster than Sherlock in typical settings, with comparable accuracy in diagnosis.


measurement and modeling of computer systems | 2012

On the efficacy of fine-grained traffic splitting protocols in data center networks

Advait Dixit; Pawan Prakash; Ramana Rao Kompella; Y. Charlie Hu

Multi-rooted tree topologies are commonly used to construct high-bandwidth data center network fabrics. In these networks, switches typically rely on equal-cost multipath (ECMP) routing techniques to split traffic across multiple paths, where each flow is routed through one of the available paths, but packets within a flow traverse the same end-to-end path. Unfortunately, since ECMP splits traffic based on flow-granularity, it can cause load imbalance across multiple paths resulting in poor utilization of network resources. More fine-grained traffic splitting techniques are typically not preferred because they can cause packet reordering that can, according to conventional wisdom, lead to severe TCP throughput degradation. In this paper, we revisit this fact in the context of regular data center topologies such as fat-tree architectures. We argue that packet-level traffic splitting, where packets belong to a given flow are sprayed through all available paths, would lead to a better load-balanced network, which in turn leads to significantly more balanced queues and much higher throughput compared to ECMP. We conduct extensive simulations to corroborate this claim.


acm special interest group on data communication | 2012

Report on the SIGCOMM 2011 conference

John W. Byers; Jeffrey C. Mogul; Fadel Adib; Jay Aikat; Danai Chasaki; Ming-Hung Chen; Marshini Chetty; Romain Fontugne; Vijay Gabale; László Gyarmati; Katrina LaCurts; Qi Liao; Marc Mendonca; Trang Cao Minh; S.H. Shah Newaz; Pawan Prakash; Yan Shvartzshnaider; Praveen Yalagandula; Chun-Yu Yang

This document provides reports on the presentations at the SIGCOMM 2011 Conference, the annual conference of the ACM Special Interest Group on Data Communication (SIGCOMM).


acm ifip usenix international conference on middleware | 2010

dFault: fault localization in large-scale peer-to-peer systems

Pawan Prakash; Ramana Rao Kompella; Venugopalan Ramasubramanian; Ranveer Chandra

Distributed hash tables (DHTs) have been adopted as a building block for large-scale distributed systems. The upshot of this success is that their robust operation is even more important as mission-critical applications begin to be layered on them. Even though DHTs can detect and heal around unresponsive hosts and disconnected links, several hidden faults and performance bottlenecks go undetected, resulting in unanswered queries and delayed responses. In this paper, we propose dFault, a system that helps large-scale DHTs to localize such faults. Informed with a log of failed queries called symptoms and some available information about the hosts in the DHT, dFault identifies the potential root causes (hosts and overlay links) that with high likelihood contributed towards those symptoms. Its design is based on the recently proposed dependency graph modeling and inference approach for fault localization. We describe the design of dFault, and show that it can accurately localize the root causes of faults with modest amount of information collected from individual nodes using a real prototype deployed over PlanetLab.


networked systems design and implementation | 2012

The TCP outcast problem: exposing unfairness in data center networks

Pawan Prakash; Advait Dixit; Y. Charlie Hu; Ramana Rao Kompella


Archive | 2013

Jumbo Frames or Not: That is the Question!

Pawan Prakash; Myungjin Lee; Y. Charlie Hu; Ramana Rao Kompella


Archive | 2013

Impact of network protocols on data center applications

Ramana Rao Kompella; Y. C. Hu; Pawan Prakash

Collaboration


Dive into the Pawan Prakash's collaboration.

Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar
Top Co-Authors

Avatar

Fadel Adib

Massachusetts Institute of Technology

View shared research outputs
Top Co-Authors

Avatar

Jay Aikat

University of North Carolina at Chapel Hill

View shared research outputs
Top Co-Authors

Avatar
Researchain Logo
Decentralizing Knowledge