The Impact of Algorithmic Trading


Published on

Studying the positive and negative impacts of algo trading on the markets.

1 Like
  • Be the first to comment

No Downloads
Total views
On SlideShare
From Embeds
Number of Embeds
Embeds 0
No embeds

No notes for slide

The Impact of Algorithmic Trading

  1. 1. The Impact of Algorithmic TradingAtharv Tilak, Priyanshi Kejriwal, Siddharth Venkataraman, LovLoothra, Parth Krishnan Mannadiar, Ritu Kapse and Puneeth C Faculty Mentor Dr. Ashish Varma
  2. 2. BackgroundThe stock market is a place where you can market or trade a company’s stock, which thecorporation issues through shares in order to raise capital. The people who buy these sharesare shareholders, and the term can refer to an individual or an organization.The stock market involves trading of bonds, which is a debt security that stipulates that theissuer of the bonds holds the holders a debt. It is exactly like a loan only that it is in form of asecurity. These bonds are traded over-the-counter, which means they are traded between thetwo parties.Instead of promising your money back, companies give a share of ownership. If there are amillion shares and you own 1000, you own 0.01% of the company.The stock market is different from the stock exchange, which is primarily concerned withbringing together buyers and sellers of stocks and securities.There are two types of exchanges where stocks can be traded. There is the exchange that hasa physical location where verbal trading takes place. The other type of exchange is the virtualkind where traders deal electronically through computer terminals.Computerization of the order flow in financial markets began in the early 1970s, with somelandmarks being the introduction of the New York Stock Exchange’s “designated orderturnaround” system (DOT, and later SuperDOT), which routed orders electronically to theproper trading post, which executed them manually. The "opening automated reportingsystem" (OARS) aided the specialist in determining the market clearing opening price (SOR;Smart Order Routing).Program trading is defined by the New York Stock Exchange as an order to buy or sell 15 ormore stocks valued at over US$1 million total. In practice this means that all program tradesare entered with the aid of a computer. In the 1980s program trading became widely used intrading between the S&P500 equity and futures markets.In stock index arbitrage a trader buys (or sells) a stock index futures contract such as the S&P500 futures and sells (or buys) a portfolio of up to 500 stocks (can be a much smallerrepresentative subset) at the NYSE matched against the futures trade. The program trade atthe NYSE would be pre-programmed into a computer to enter the order automatically into theNYSE’s electronic order routing system at a time when the futures price and the stock indexwere far enough apart to make a profit.At about the same time portfolio insurance was designed to create a synthetic put option on astock portfolio by dynamically trading stock index futures according to a computer modelbased on the Black–Scholes option pricing model.Both strategies, often simply lumped together as "program trading", were blamed by manypeople (for example by the Brady report) for exacerbating or even starting the 1987 stock
  3. 3. market crash. Yet the impact of computer driven trading on stock market crashes is unclearand widely discussed in the academic community.Financial markets with fully electronic execution and similar electronic communicationnetworks developed in the late 1980s and 1990s. In the U.S., decimalization, which changedthe minimum tick size from 1/16 of a dollar (US$0.0625) to US$0.01 per share, may haveencouraged algorithmic trading as it changed the market microstructure by permitting smallerdifferences between the bid and offer prices, decreasing the market-makers tradingadvantage, thus increasing market liquidity.This increased market liquidity led to institutional traders splitting up orders according tocomputer algorithms so they could execute orders at a better average price. These averageprice benchmarks are measured and calculated by computers by applying the time-weightedaverage price or more usually by the volume-weighted average price.A further encouragement for the adoption of algorithmic trading in the financial marketscame in 2001 when a team of IBM researchers published a paper at the International JointConference on Artificial Intelligence where they showed that in experimental laboratoryversions of the electronic auctions used in the financial markets, two algorithmic strategies(IBMs own MGD, and Hewlett-Packards ZIP) could consistently out-perform humantraders. MGD was a modified version of the "GD" algorithm invented by Steven Gjerstad &John Dickhaut in 1996/7; the ZIP algorithm had been invented at HP by Dave Cliff(professor) in 1996. In their paper, the IBM team wrote that the financial impact of theirresults showing MGD and ZIP outperforming human traders "...might be measured in billionsof dollars annually"; the IBM paper generated international media coverage.As more electronic markets opened, other algorithmic trading strategies were introduced.These strategies are more easily implemented by computers, because machines can reactmore rapidly to temporary mispricing and examine prices from several marketssimultaneously. For example Stealth (developed by the Deutsche Bank), Sniper and Guerilla(developed by Credit Suisse), arbitrage, statistical arbitrage, trend following, and meanreversion.This type of trading is what is driving the new demand for Low Latency Proximity Hostingand Global Exchange Connectivity. It is imperative to understand latency when puttingtogether a strategy for electronic trading. Latency refers to the delay between the transmissionof information from a source and the reception of the information at a destination. Latencyhas as a lower bound the speed of light; this corresponds to about 3.3 milliseconds per 1,000kilometers of optical fibre. Any signal regenerating or routing equipment introduces greaterlatency than this speed-of-light baseline.
  4. 4. IntroductionInvestment decisions have been historically made by employing one of two strategic actions.One is based upon the economic fundamentals of the company issuing the equity and theother on the behavioral patterns of the buyers and sellers of the said equity.Until recently, it was almost impossible to act upon the technical anomalies in a timelymanner and at a large enough scale to create tangible benefits. The digitization of tradingactivities has fundamentally altered the landscape of financial markets. A research conductedby the Aite Group revealed that in 2006, a third of all European Union and United Statesstock trades were driven by algorithms. In 2009, the Tabb Group estimated that HighFrequency Trading firms accounted for as much as 73% of all US equity trading volume.Moreover, On May 6, 2010 U.S. stock market indices, stock-index futures, options, andexchange-traded funds experienced a sudden price drop of more than five percent, followedby a rapid rebound. This brief period of extreme intraday volatility, commonly referred to asthe “Flash Crash”, raises a number of questions about the structure and stability of markets. A survey conducted by Market Strategies International in June 2010revealed that over 80 percent of U.S. retail advisors believe that “overreliance on computersystems and high-frequency trading” were the primary contributors to the volatility observedon May 6.It is evident that with such a widespread adoption and use, these algorithms impact companystrategies, investment practices and the very markets in which they’re deployed. The study ofthe various forms of algorithmic trading and their impact on the market forms the basis forthis paper.What is algorithmic trading?Algorithmic trading (AT) is any trading activity carried out with the help of algorithms. It canbe formally defined as “placing a buy or sell order of a defined quantity into a quantitativemodel that automatically generates the timing of orders and the size of orders based on goalsspecified by the parameters and constraints of the algorithm”.The rules built into the model attempt to determine the optimal time for an order to be placedthat will cause the least amount of impact on the price of the financial instrument.Algorithmic trading is a way to codify a trader’s execution strategy and it also cuts downtransaction costs and allows fund managers to take control of their own trading processes.Algorithmic trading is widely used by pension funds, mutual funds, and other buy side(investor driven) institutional traders, to divide large trades into several smaller trades tomanage market impact, and risk. Buy-side firms are gravitating toward rules-based systems.For example, instead of placing 1,00,000-share order, an algorithmic trading strategy maypush 1,000 shares out every 30 seconds and incrementally feed small amounts into the market
  5. 5. over the course of several hours or the entire day. By breaking their large orders into smallerchunks, buy-side institutions are able to disguise their orders and participate in a stock’strading volume across an entire day or for a few hours. The time frame depends on thetraders’ objective, how aggressive they want to be, and constraints such as size, price, andorder type, liquidity and volatility of the stock and industry group. More sophisticatedalgorithms allow buy-side firms to fine-tune the trading parameters in terms of start time, endtime, and aggressiveness. Algorithmic trading is appealing to buy-side firms because they canmeasure their trading results against industry-standard benchmarks such as volume weightedaverage price (VWAP) or the S&P 500 and Russell 3000 indices.Sell side traders, such as market makers and some hedge funds, provide liquidity to themarket, generating and executing orders automatically. In their never-ending quest to pleasetheir customers, being the first to innovate can give a broker a significant advantage over thecompetition, both in capturing the order flow of early adopters and building a reputation as athought leader. It is possible to create an algorithm and enjoy a significant time windowahead of the competition if that algorithm addresses a really unique execution strategy.A special class of algorithmic trading is "high-frequency trading" (HFT), in which computersmake elaborate decisions to initiate orders based on information that is receivedelectronically, before human traders are capable of processing the information they observe.This has resulted in a dramatic change of the market microstructure, particularly in the wayliquidity is provided.Algorithmic trading may be used in any investment strategy, including market making, inter-market spreading, arbitrage, or pure speculation (including trend following). The investmentdecision and implementation may be augmented at any stage with algorithmic support or mayoperate completely automatically.Evolution of Algorithmic tradingNearly 40 years ago, when exchanges first contemplated switching to fully automated tradingplatforms, Fischer Black surmised that regardless of market structure, liquid markets exhibitprice continuity only if trading is characterized by a large volume of small individual trades.Large order executions would always exert an impact on price, irrespective of the method ofexecution or technological advances in market structure.At that time, stock market “specialists” were officially designated market makers, obligatedto maintain the order book and provide liquidity. In the trading pits of the futures markets,many floor traders were unofficial, but easily identifiable market makers. Both the stockmarket specialists and futures market floor traders enjoyed a proximity advantage comparedto traders who participated away from the trading floor. This advantage allowed specialistsand floor traders to react more quickly to incoming order flow compared to other traders.
  6. 6. As markets became electronic, a rigid distinction between market makers and other tradersbecame obsolete. Securities exchanges increasingly adopted a limit order market design, inwhich traders submit orders directly into the exchange’s electronic systems, bypassing bothdesignated and unofficial market makers. In today’s electronic markets, High FrequencyTraders enjoy a latency advantage which allows them to react to changes in order flow morequickly than other traders. This occurred because of advances in technology, as well asregulatory requirements and also because large institutional investors have been allowed toplace their servers in the same building as the stock exchanges, and they are able to gleam attransactions a few milliseconds before the other market players do.Types of algorithmic tradesThe algorithms employed in automated trading can be broadly categorized into two“families” viz. Execution Algorithms and Alpha-Generating Algorithms.Execution Algorithms: These programs execute stock market trades in such a manner thatthe prices aren’t influenced by momentary swings in the market. Two of the commonexecution algorithms are the Volume Weighted Average Price (VWAP) and Time WeightedAverage Price (TWAP).The VWAP is calculated by weighting a stock’s price quotes through the trading session withvolumes traded at each price. The algorithm’s objective is to execute the order at a price thatis as close as possible to this weighted average. The reason why this algorithm is useful canbe gleaned from the following example. Let us say that Stock A traded in a range of Rs. 1090to Rs. 1,110 on a particular day and its VWAP was Rs. 1,100. An investor selling the stock atthe VWAP instead of at the day’s low, would have gained as much as Rs. 10 per share.Institutional investors which measure their returns based on end-of-day prices may also usethe VWAP for the last 30 minutes to buy or sell their holdings to reduce deviations.The TWAP strategy simply breaks up a large order into equal parts and then dribbles buy orsell orders into the market evenly over the trading session. This basic strategy is called"iceberging". This ensures that the price at which the investor buys or sells is not distorted bymomentary blips in the market. Using a TWAP is much like using a mutual fund systematicinvestment plan (SIP) — only compressed into minutes rather than months.Alpha Generating/Seeking Algorithms: Unlike execution algorithms, alpha-seekingalgorithms actively try to make money. They track historical relationships between securities,assets or markets and then exploit minor deviations for quick gains. A few of the commonalpha seeking strategies are given below:Arbitrage algorithms: When used by academics, an arbitrage is a transaction that involves nonegative cash flow at any probabilistic or temporal state and a positive cash flow in at leastone state; in simple terms, it is the possibility of a risk-free profit at zero cost. Therefore,much like the traditional arbitrageurs do, arbitrage algorithms earn a spread from trading on
  7. 7. anomalies between securities, trading venues or asset classes. For example, simple arbitragealgorithms may earn a ‘spread’ by buying a stock at Rs. 100 on the BSE and selling it at Rs.100.50 on the NSE. Arbitrage profits can also be earned by exploiting differentials betweenfutures and cash markets.Arbitrage is possible when one of three conditions is met: 1. The same asset does not trade at the same price on all markets (the "law of one price"). 2. Two assets with identical cash flows do not trade at the same price. 3. An asset with a known price in the future does not today trade at its future price discounted at the risk-free interest rate (or, the asset does not have negligible costs of storage; as such, for example, this condition holds for grain but not for securities).Arbitrage is not simply the act of buying a product in one market and selling it in another fora higher price at some later time. The transactions must occur simultaneously to avoidexposure to market risk, or the risk that prices may change on one market before bothtransactions are complete.In practical terms, this is generally only possible with securities and financial products whichcan be traded electronically, and even then, when each leg of the trade is executed the pricesin the market may have moved. Missing one of the legs of the trade (and subsequently havingto trade it soon after at a worse price) is called execution risk or more specifically leg risk.In the simplest example, any good sold in one market should sell for the same price inanother. Traders may, for example, find that the price of wheat is lower in agriculturalregions than in cities, purchase the good, and transport it to another region to sell at a higherprice. This type of price arbitrage is the most common, but this simple example ignores thecost of transport, storage, risk, and other factors. "True" arbitrage requires that there be nomarket risk involved. Where securities are traded on more than one exchange, arbitrageoccurs by simultaneously buying in one and selling on the other.A variation is the ‘event’ arbitrage, which exploits money-making opportunities arising frommergers, buyouts or restructuring. An algorithmic trader, alerted instantly to an acquisition,could buy the acquirer’s stock and short the target’s.Trend following algorithms: Trend-following algorithms techniques commonly used bytechnical analysts to identify a reversal in trends. They then piggyback on it at an early stageto benefit from the momentum. These algorithms may track technical indicators such as the50 or 200-day moving averages or relative strength index, to bet on stocks on the verge ofbreaking out or breaking down.Traders who subscribe to a trend following strategy do not aim to forecast or predict specificprice levels; they initiate a trade when a trend appears to have started, and exit the trade oncethe trend appears to have ended.
  8. 8. Scalping: Scalping is a method of arbitrage of small price gaps created by the bid-ask spread.Scalpers attempt to act like traditional market makers or specialists. To make the spreadmeans to buy at the bid price and sell at the ask price, to gain the bid/ask difference. Thisprocedure allows for profit even when the bid and ask do not move at all, as long as there aretraders who are willing to take market prices. It normally involves establishing andliquidating a position quickly, usually within minutes or even seconds.The role of a scalper is actually the role of market makers or specialists who are to maintainthe liquidity and order flow of a product of a market. A market maker is basically aspecialized scalper. The volume a market maker trades are many times more than the averageindividual scalpers. A market maker has a sophisticated trading system to monitor tradingactivity. However, a market maker is bound by strict exchange rules while the individualtrader is not. For instance, NASDAQ requires each market maker to post at least one bid andone ask at some price level, so as to maintain a two-sided market for each stock represented.Double hedging: Double Hedging describes a strategy that supervises pairs of assets withidentical statistical correlations. Changes in market price of the given pair of assets aretracked, and once the change occurs the algorithm counts the difference in standard deviationbetween the two assets, and then executes the hedging procedure. For example, “Shell” and“Exxon” shares are both in the energy sector and therefore are correlated. Following this pairof stocks allows the algorithm to determine if future profits can be made, and the possibilityof profit increases as the correlation between the two assets gets smaller. Thus, “doublehedging” occurs as the correlation value goes down, presenting the opportunity tosimultaneously buy one share and sell the other thereby locking in a profit.High Frequency TradingThe very essence of algorithmic trading lies in the fact that the machines running thesealgorithms are capable of processing a huge volume of information in a relatively shortperiod of time. This particular fact is exploited in High-frequency trading (HFT) which is aspecial class of algorithmic trading in which computers make elaborate decisions to initiateorders based on information that is received electronically, before human traders are capableof processing the information they observe. This has resulted in a dramatic change of themarket microstructure, particularly in the way liquidity is provided.HFT is typically characterized by several distinguishing features:  It is highly quantitative, employing computerized algorithms to analyze incoming market data and implement proprietary trading strategies  An investment position is held only for very brief periods of time - from seconds to hours - and rapidly trades into and out of those positions, sometimes thousands or tens of thousands of times a day  At the end of a trading day there is no net investment position;
  9. 9.  It is mostly employed by proprietary firms or on proprietary trading desks in larger, diversified firms  It is very sensitive to the processing speed of markets and of their own access to the market  Many high-frequency traders provide liquidity and price discovery to the markets through market-making and arbitrage trading; high-frequency traders also take liquidity to manage risk or lock in profitsHigh-frequency traders compete on a basis of speed with other high-frequency traders, notlong-term investors (who typically look for opportunities over a period of weeks, months, oryears), and compete for very small, consistent profits. As a result, high-frequency trading hasbeen shown to have a potential Sharpe ratio (measure of reward per unit of risk) thousands oftimes higher than the traditional buy-and-hold strategies.Aiming to capture just a fraction of a penny per share or currency unit on every trade, high-frequency traders move in and out of such short-term positions several times each day.Fractions of a penny accumulate fast to produce significantly positive results at the end ofevery day. High-frequency trading firms do not employ significant leverage, do notaccumulate positions, and typically liquidate their entire portfolios on a daily basis.Recently, HFT has become more prominent and controversial. These algorithms ortechniques are commonly given names such as "Stealth" (developed by the Deutsche Bank),"Iceberg", "Dagger", "Guerrilla", "Sniper", "BASOR" (developed by Quod Financial) and"Sniffer". Yet are quite simple mathematical constructs at their core. Dark pools arealternative electronic stock exchanges where trading takes place anonymously, with mostorders hidden or "iceberged." Gamers or "sharks" sniff out large orders by "pinging" smallmarket orders to buy and sell. When several small orders are filled the sharks may havediscovered the presence of a large iceberged order.“Now it’s an arms race,” said Andrew Lo, director of the Massachusetts Institute ofTechnology’s Laboratory for Financial Engineering. “Everyone is building moresophisticated algorithms, and the more competition exists, the smaller the profits.”Analyzing the ImpactOne of the major criticisms leveled at this form of trading is that from the point of view ofhow capital is allocated, they are inherently self-referential. That is, they represent thefinancial markets looking inward at their own dynamics and investing capital to create morecapital without ever attending to the externalities of this behavior. Therefore, as long as thealgorithm trading account creates sufficient risk-adjusted returns a financial manager hasnothing to worry about. There seems to be no reason for anyone to embrace the ambiguitiesof fundamental investing if one could generate returns in this more direct way.
  10. 10. The result of algorithmic trading’s success, predictably, has been that more and more capitalhas been attracted to this investment focus. The resulting increase in crowdedness has meantthat the anomalies now available for exploitation are both much more short-lived (hencedriving the aforementioned “arms race” to micro-second execution windows) and also muchmore granular, meaning that more capital has to be applied during these windows to creatematerial results, in turn driving the need for higher and higher leverage ratios.Market Volatility: The consequence of these developments has been an escalation in marketvolatility, an increase in unhedgeable catastrophic risk, and thus a decrease in risk-adjustedreturns (if one truly accounted for catastrophic risk), all resulting in an unsustainableoverweight in the overall market’s portfolio away from fundamental investing into short-termtrading. Worst of all, this volatility has made it much harder for fundamental investors toenter the market. An article in the New York Times had an example of how HFT affectsshare prices and they used a $1.4 million order on which they said the slow moving investorshad to pay $7,800 extra because of high frequency traders which is about half a percent moreon the trade than they would have otherwise had to.The Stock Market Crash of 1987: It has been said that the October 1987 stock market crashwas caused in part by something called dynamic portfolio insurance, another approach basedon algorithms. Dynamic portfolio insurance is a way of protecting your portfolio of shares sothat if the market falls you can limit your losses to an amount you stipulate in advance. As themarket falls, you sell some shares. By the time the market falls by a certain amount, you willhave closed all your positions so that you can lose no more money.A proper implementation of the idea required some knowledge of option theory as developedby the economists Fischer Black of Goldman Sachs, Myron S. Scholes of Stanford andRobert C. Merton of Harvard. You type into some formula the current stock price, and thistells you how many shares to hold. The market falls and you type the new price into theformula, which tells you how many to sell.By 1987, however, the problem was the sheer number of people following the strategy andthe market share that they collectively controlled. If a fall in the market leads to peopleselling according to some formula, and if there are enough of these people following the samealgorithm, then it will lead to a further fall in the market, and a further wave of selling, and soon — until the S&P 500 index loses over 20 percent of its value in single day: Oct. 19, BlackMonday. Dynamic portfolio insurance caused the very thing it was designed to protectagainst.This is the sort of feedback that occurs between a popular strategy and the underlying market,with a long-lasting effect on the broader economy. A rise in price begets a rise and a fallbegets a fall and as a result, volatility rises and the market is destabilized. All that’s needed isfor a large number of people to be following the same type of strategy.The “Flash Crash” of 2010: As had been stated at the beginning of this article, most U.S.retail advisors believe that HFT had a major part to play in the 2010 “Flash Crash”. A paperby Kirilenko, Kyle, Samadi and Tuzun uses audit-trail data to describe events that took place
  11. 11. on that day and how the HFT responses to the unusually large selling pressure exacerbatedmarket volatility.According to the paper, the events on May 6 unfolded as follows:Financial markets, already tense over concerns about the European sovereign debt crisis,opened to news concerning the Greek government’s ability to service its sovereign debt. As aresult, premiums rose for buying protection against default on sovereign debt securities ofGreece and a number of other European countries. In addition, the S&P 500 volatility index(“VIX”) increased, and yields of ten-year Treasuries fell as investors engaged in a “flight toquality.” By mid-afternoon, the Dow Jones Industrial Average was down about 2.5%.Sometime after 2:30 p.m., Fundamental Sellers began executing a large sell program.Typically, such a large sell program would not be executed at once, but rather spread out overtime, perhaps over hours. The magnitude of the Fundamental Sellers’ trading program beganto significantly outweigh the ability of Fundamental Buyers to absorb the selling pressure.HFTs and Intermediaries were the likely buyers of the initial batch of sell orders fromFundamental Sellers, thus accumulating temporary long positions. Thus, during the earlymoments of this sell program’s execution, HFTs and Intermediaries provided liquidity to thissell order.However, just like market intermediaries in the days of floor trading, HFTs andIntermediaries had no desire to hold their positions over a long time horizon. A few minutesafter they bought the first batch of contracts sold by Fundamental Sellers, HFTs aggressivelysold contracts to reduce their inventories. As they sold contracts, HFTs were no longerproviders of liquidity to the selling program. In fact, HFTs competed for liquidity with theselling program, further amplifying the price impact of this program.Furthermore, total trading volume and trading volume of HFTs increased significantlyminutes before and during the Flash Crash. Finally, as the price of the E-mini rapidly fell andmany traders were unwilling or unable to submit orders, HFTs repeatedly bought and soldfrom one another, generating a “hot-potato” effect.Yet, Fundamental Buyers, who may have realized significant profits from this large decreasein price, did not seem to be willing or able to provide ample buy-side liquidity. As a result,between 2:45:13 and 2:45:27, prices of the E-mini fell about 1.7%.At 2:45:28, a 5 second trading pause was automatically activated in the E-mini. Opportunisticand Fundamental Buyers aggressively executed trades which led to a rapid recovery in prices.HFTs continued their strategy of rapidly buying and selling contracts, while about half of theIntermediaries closed their positions and got out of the market.In light of these events, a few fundamental questions arise. Why did it take so long forFundamental Buyers to enter the market and why did the price concessions had to be solarge? It seems possible that some Fundamental Buyers could not distinguish betweenmacroeconomic fundamentals and market-specific liquidity events. It also seems possible that
  12. 12. the opportunistic buyers have already accumulated a significant positive inventory earlier inthe day as prices were steadily declining. Furthermore, it is possible that they could notquickly find opportunities to hedge additional positive inventory in other markets which alsoexperienced significant volatility and higher latencies.The Knight Capital Case: In a recent case eerily similar to the Flash Crash of 2010, amalfunction in Knight Capital’s trading system flooded the market with erroneous trades onAugust 1, 2012. The troubles resulted in some $440 million in trading losses — nowestimated at roughly $275 million after taxes — which left the firm scrambling for a financialrescue.While that bailout came, there was a steep cost beyond just the trading losses, namely controlof the firm. Knight’s new investors control nearly three-quarters of the company, and they gotit for a song; existing shareholders have seen their stakes significantly diluted, and thecompany faces the daunting task of retaining clients.Knight is a trading firm that takes orders from big brokers like E-Trade and TD Ameritrade,routing them to the exchanges where the stocks are traded. One key role Knight Capital hasplayed is that of a designated market maker, where it is responsible for maintaining orderlytrades in the stocks it oversees.Designated market makers are particularly important whenever there is a lot of marketvolatility; Knight is responsible for the trading of 524 NYSE-listed stocks, a sizable chunk ofthe roughly 2,300 total corporate issuers.Until its blunder, Knight was a respected top-level player, a distinction that makes its slipmore alarming, because the worry is that if this could happen at Knight, it can happenanywhere.Increase in Liquidity: Although we’ve analyzed a lot of cases which reflect negatively onthe idea of Algorithmic trading, its important not to neglect some of the benefits thatalgorithmic trading brings to the table. According to a research paper on the topic byHendershott, Jones, and Menkveld, algorithmic trading “improves liquidity and enhances theinformativeness of quotes."The authors of the paper emphasize that automated trading now runs through the markets atevery level and that there are many different algorithms, used by many different types ofmarket participants. Some hedge funds and broker-dealers supply liquidity using algorithms,competing with designated market-makers and other liquidity suppliers. For assets that tradeon multiple venues, liquidity demanders often use smart order routers to determine where tosend an order. Statistical arbitrage funds use computers to quickly process large amounts ofinformation contained in the order flow and price moves in various securities, trading at highfrequency based on patterns in the data. Last but not least, algorithms are used by institutionalinvestors to trade large quantities of stock gradually over time.One very important observation made in the paper is that it is not at all obvious thatalgorithmic trading should improve market liquidity. If algorithms are cheaper and/or better
  13. 13. at supplying liquidity, then AT may result in more competition in liquidity provision, therebylowering the cost of immediacy. However, the effects could go the other way if algorithmsare used mainly to demand liquidity. Limit order submitters grant a trading option to others,and if algorithms make liquidity demanders better able to identify and pick off an in-the-money trading option, then the cost of providing the trading option increases, and spreadsmust widen to compensate. In fact, AT could actually lead to an unproductive arms race,where liquidity suppliers and liquidity demanders both invest in better algorithms to try totake advantage of the other side, with measured liquidity the unintended victim.This is the kind of thing most participants in algorithmic trading do not emphasize whenraving about the obvious benefits it brings to markets.One of the highlights of the paper is the effort to track the rise of algorithmic trading (overroughly a five year period, 2001-2006) and to compare this to changes in liquidity. This isntquite as easy as it might seem because algorithmic trading is just trading and not obviouslydistinct in market records from other trading. One cannot directly observe whether aparticular order is generated by a computer algorithm. For cost and speed reasons, mostalgorithms do not rely on human intermediaries but instead generate orders that are sentelectronically to a trading venue. Therefore, one can use the rate of electronic message trafficas a proxy for the amount of algorithmic trading taking place.The figure below shows this data, recorded for stocks with differing market capitalization(sorted into quintiles, Q1 being the largest fifth). Clearly, the amount of electronic traffic inthe trading system has increased by a factor of at least five over a period of five years:
  14. 14. The paper then compares this to data on the effective bid-ask spread for this same set ofstocks, again organized by quintile, over the same period. The resulting figure indeed shows amore or less steady decrease in the spread, a measure of improving liquidity:So, there is a clear correlation. The next question, of course, is whether this correlationreflects a causal process or not. What perhaps sets this study apart from others (see, forexample, any number of reports by the Tabb Group, which monitors high-frequency markets)is an effort to get at this causal link. The authors do this by studying a particular historicalevent that increased the amount of algorithmic trading in some stocks but not others. Theresults suggest that there is a causal link.The conclusion, then, is that algorithmic trading (at least in the time period studied, in whichstocks were generally rising) does improve market efficiency in the sense of higher liquidityand better price discovery. But there is a further caveat. While the authors have accounted forshare price levels and volatility in their study, it remains an open question whetheralgorithmic trading and algorithmic liquidity supply are equally beneficial in more turbulentor declining markets. Like Nasdaq market makers refusing to answer their phones during the1987 stock market crash, algorithmic liquidity suppliers may simply turn off their machineswhen markets spike downward.This resonates with a general theme across all finance and economics. When markets arebehaving "normally", they seem to be more or less efficient and stable. When they gohaywire, all the standard theories and accepted truths go out the window. Unfortunately,"haywire" isnt as unusual as many theorists would like it to be.
  15. 15. Expert ViewWe were honored to have had an opportunity to interview Mr. Sanket Kapse, an analyst in alarge Hedge Fund. We gained the following insights from the interview:Please tell us a bit about your company, profile and designation.I work at DE Shaw and Company as a Finance and Operations Generalist. My profile is thatof a Back Office Analyst.Could you elaborate on the nature of your work at DE Shaw?I take care of the back office functions for the Algorithmic trading (AT) portfolio whichincludes PNL generation, corporate actions processing, performing various reconciliationsand end of day management reporting (MIS).Are you directly involved in Algorithm Trading activities of your company?There is no direct role of AT in my work but my firm uses the same extensively for trading invarious geographies including Indian markets.Can you please throw some light on the subject?AT put simply, is using the power of computers to implement certain mathematical models todiscover anomalies (arbitrage opportunities) in the prices of financial instruments to make aquick return. A simple example would be trading bonds. Bonds can be valued usingparameters like current market interest rates, coupon rate of the bond and the risk profile ofthe bond issuer. Computers can be fed a mathematical model and made to value hundreds ofbonds each day to spot opportunities where a bond is either trading cheap or expensive. Thealgorithms can also trade in and out of the bond to make a quick profit. Same can be done fortrading stocks, commodities, currencies, interest rates etc.In your opinion, is it ethical to employ AT?AT is just a way of trading which is faster and smarter than how human traders can trade. Italso needs a lot of investment into technology and requires hiring math and science expertswhich can be only done by players who have deep pockets. It thus puts the less privilegedtraders at a disadvantage.AT is also used by large market players like hedge funds and investment banks for better riskmanagement as they trade in and out of a position very quickly and hence take-on very littlemarket related risk.For the above two reasons many have termed AT as un-ethical. However, on the positiveside, it adds a lot of depth and liquidity to the market. 1. Since AT is done on a high frequency, it creates better liquidity in the markets. This ensures better price discovery and prevents huge volatility in the prices that can be caused by scanty liquidity. 2. The high volumes generated by AT keeps the BID-ASK spreads very low. For e.g. if there is only one foreign exchange shop in your city, a tourist would have to buy dollars at say Rs. 50 per dollar for his trip. But on returning any residual dollars would not fetch more than Rs. 45. The spread here is Rs. 5 which is a loss to the tourist this would not have happened if there were 100 foreign exchange shops.
  16. 16. 3. AT is a battle between sophisticated players that operate in the ultra-short term time frame. It does not negatively impact long-term investors as they do not compete for long-term returns.Recent developmentsFinancial market news is now being formatted by firms such as Need To Know News,Thomson Reuters, Dow Jones, and Bloomberg, to be read and traded on via algorithms."Computers are now being used to generate news stories about company earnings results oreconomic statistics as they are released. And this almost instantaneous information forms adirect feed into other computers which trade on the news."The algorithms do not simply trade on simple news stories but also interpret more difficult tounderstand news. Some firms are also attempting to automatically assign sentiment (decidingif the news is good or bad) to news stories so that automated trading can work directly on thenews story."Increasingly, people are looking at all forms of news and building their own indicatorsaround it in a semi-structured way," as they constantly seek out new trading advantages saidRob Passarella, global director of strategy at Dow Jones Enterprise Media Group. His firmprovides both a low latency news feed and news analytics for traders. Passarella also pointedto new academic research being conducted on the degree to which frequent Google searcheson various stocks can serve as trading indicators, the potential impact of various phrases andwords that may appear in Securities and Exchange Commission statements and the latestwave of online communities devoted to stock trading topics."Markets are by their very nature conversations, having grown out of coffee houses andtaverns", he said. So the way conversations get created in a digital society will be used toconvert news into trades, as well, Passarella said.“There is a real interest in moving the process of interpreting news from the humans to themachines” says Kirsti Suutari, global business manager of algorithmic trading at Reuters."More of our customers are finding ways to use news content to make money."An example of the importance of news reporting speed to algorithmic traders was anadvertising campaign by Dow Jones (appearances included page W15 of the Wall StreetJournal, on March 1, 2008) claiming that their service had beaten other news services by 2seconds in reporting an interest rate cut by the Bank of England.In July 2007, Citigroup, which had already developed its own trading algorithms, paid $680million for Automated Trading Desk, a 19-year-old firm that trades about 200 million sharesa day. Citigroup had previously bought Lava Trading and OnTrade Inc.In late 2010, The UK Government Office for Science initiated a Foresight projectinvestigating the future of computer trading in the financial markets, led by Dame Clara
  17. 17. Furse, ex-CEO of the London Stock Exchange and in September 2011 the project publishedits initial findings in the form of a three-chapter working paper available in three languages,along with 16 additional papers that provide supporting evidence. All of these findings areauthored or co-authored by leading academics and practitioners, and were subjected toanonymous peer-review. The Foresight project is set to conclude in late 2012.In September 2011, RYBN has launched "ADM8", an open source Trading Bot prototype,already active on the financial marketsReferences 1. Terrence Hendershott, Charles M. Jones, and Albert J. Menkveld, “Does Algorithmic Trading Improve Liquidity?,” The Journal of Finance, Vol. LXVI, No. 1, 2011 2. Andrei Kirilenko, Mehrdad Samadi, Albert S. Kyle, Tugkan Tuzun, “The Flash Crash: The Impact of High Frequency Trading on an Electronic Market,” Available at SSRN:, 2011 3. Markus Gsell, “Assessing the Impact of Algorithmic Trading on Markets: A Simulation Approach,” Available at SSRN: 4. Alain Chaboud, Benjamin Chiquoine, Erik Hjalmarsson, Clara Vega, “Rise of the Machines: Algorithmic Trading in the Foreign Exchange Market,” Available at SSRN: 5. Alvaro Carteay and Sebastian Jaimungal, “Modeling Asset Prices for Algorithmic and High Frequency Trading,” Available at SSRN: 6. S. Gjerstad and J. Dickhaut, "Price Formation in Double Auctions,” Games and Economic Behavior, Volume 22, Issue 1, p. 1, January, 1998 7. Michael Mackenzie, “SEC runs eye over high-speed trading,” The Financial Times, July 29, 2010, p. 21 8. Aline van Duyn, "If youre reading this, its too late: a machine got here first," The Financial Times, April 16, 2007, p.1 9. Charles Duhigg, “Stock Traders Find Speed Pays, in Milliseconds,” The New York Times, July 23, 2009 10. Paul Wilmott, “Hurrying Into the Next Panic?,” The New York Times, July 28, 2009 11. Caroline Vatetkevitch, Chuck Mikolajczak, "Error by Knight Capital rips through stock market,” Reuters, August 1, 2012