In statistics, the k-nearest neighbors algorithm (k-NN) is a non-parametricsupervised learning method. It was first developed by Evelyn Fix and Joseph Hodges in 1951,[1] and later expanded by Thomas Cover.[2]
Most often, it is used for classification, as a k-NN classifier, the output of which is a class membership. An object is classified by a plurality vote of its neighbors, with the object being assigned to the class most common among its k nearest neighbors (k is a positive integer, typically small). If k = 1, then the object is simply assigned to the class of that single nearest neighbor.
The k-NN algorithm can also be generalized for regression. In k-NN regression, also known as nearest neighbor smoothing, the output is the property value for the object. This value is the average of the values of k nearest neighbors. If k = 1, then the output is simply assigned to the value of that single nearest neighbor, also known as nearest neighbor interpolation.
For both classification and regression, a useful technique can be to assign weights to the contributions of the neighbors, so that nearer neighbors contribute more to the average than distant ones. For example, a common weighting scheme consists of giving each neighbor a weight of 1/d, where d is the distance to the neighbor.[3]
The input consists of the k closest training examples in a data set.
The neighbors are taken from a set of objects for which the class (for k-NN classification) or the object property value (for k-NN regression) is known. This can be thought of as the training set for the algorithm, though no explicit training step is required.
A peculiarity (sometimes even a disadvantage) of the k-NN algorithm is its sensitivity to the local structure of the data.
In k-NN classification the function is only approximated locally and all computation is deferred until function evaluation. Since this algorithm relies on distance, if the features represent different physical units or come in vastly different scales, then feature-wise normalizing of the training data can greatly improve its accuracy.[4]
Statistical setting
Suppose we have pairs taking values in , where Y is the class label of X, so that for (and probability distributions ). Given some norm on and a point , let be a reordering of the training data such that .
Algorithm
The training examples are vectors in a multidimensional feature space, each with a class label. The training phase of the algorithm consists only of storing the feature vectors and class labels of the training samples.
In the classification phase, k is a user-defined constant, and an unlabeled vector (a query or test point) is classified by assigning the label which is most frequent among the k training samples nearest to that query point.
A commonly used distance metric for continuous variables is Euclidean distance. For discrete variables, such as for text classification, another metric can be used, such as the overlap metric (or Hamming distance). In the context of gene expression microarray data, for example, k-NN has been employed with correlation coefficients, such as Pearson and Spearman, as a metric.[5] Often, the classification accuracy of k-NN can be improved significantly if the distance metric is learned with specialized algorithms such as Large Margin Nearest Neighbor or Neighbourhood components analysis.
A drawback of the basic "majority voting" classification occurs when the class distribution is skewed. That is, examples of a more frequent class tend to dominate the prediction of the new example, because they tend to be common among the k nearest neighbors due to their large number.[6] One way to overcome this problem is to weight the classification, taking into account the distance from the test point to each of its k nearest neighbors. The class (or value, in regression problems) of each of the k nearest points is multiplied by a weight proportional to the inverse of the distance from that point to the test point. Another way to overcome skew is by abstraction in data representation. For example, in a self-organizing map (SOM), each node is a representative (a center) of a cluster of similar points, regardless of their density in the original training data. K-NN can then be applied to the SOM.
Parameter selection
The best choice of k depends upon the data; generally, larger values of k reduces effect of the noise on the classification,[7] but make boundaries between classes less distinct. A good k can be selected by various heuristic techniques (see hyperparameter optimization). The special case where the class is predicted to be the class of the closest training sample (i.e. when k = 1) is called the nearest neighbor algorithm.
The accuracy of the k-NN algorithm can be severely degraded by the presence of noisy or irrelevant features, or if the feature scales are not consistent with their importance. Much research effort has been put into selecting or scaling features to improve classification. A particularly popular[citation needed] approach is the use of evolutionary algorithms to optimize feature scaling.[8] Another popular approach is to scale features by the mutual information of the training data with the training classes.[citation needed]
In binary (two class) classification problems, it is helpful to choose k to be an odd number as this avoids tied votes. One popular way of choosing the empirically optimal k in this setting is via bootstrap method.[9]
The 1-nearest neighbor classifier
The most intuitive nearest neighbour type classifier is the one nearest neighbour classifier that assigns a point x to the class of its closest neighbour in the feature space, that is .
As the size of training data set approaches infinity, the one nearest neighbour classifier guarantees an error rate of no worse than twice the Bayes error rate (the minimum achievable error rate given the distribution of the data).
The weighted nearest neighbour classifier
The k-nearest neighbour classifier can be viewed as assigning the k nearest neighbours a weight and all others 0 weight. This can be generalised to weighted nearest neighbour classifiers. That is, where the ith nearest neighbour is assigned a weight , with . An analogous result on the strong consistency of weighted nearest neighbour classifiers also holds.[10]
Let denote the weighted nearest classifier with weights . Subject to regularity conditions, which in asymptotic theory are conditional variables which require assumptions to differentiate among parameters with some criteria. On the class distributions the excess risk has the following asymptotic expansion[11]
for constants and where and .
The optimal weighting scheme , that balances the two terms in the display above, is given as follows: set ,
for and
for .
With optimal weights the dominant term in the asymptotic expansion of the excess risk is . Similar results are true when using a bagged nearest neighbour classifier.
The naive version of the algorithm is easy to implement by computing the distances from the test example to all stored examples, but it is computationally intensive for large training sets. Using an approximate nearest neighbor search algorithm makes k-NN computationally tractable even for large data sets. Many nearest neighbor search algorithms have been proposed over the years; these generally seek to reduce the number of distance evaluations actually performed.
k-NN has some strong consistency results. As the amount of data approaches infinity, the two-class k-NN algorithm is guaranteed to yield an error rate no worse than twice the Bayes error rate (the minimum achievable error rate given the distribution of the data).[2] Various improvements to the k-NN speed are possible by using proximity graphs.[14]
For multi-class k-NN classification, Cover and Hart (1967) prove an upper bound error rate of
where is the Bayes error rate (which is the minimal error rate possible), is the asymptotic k-NN error rate, and M is the number of classes in the problem. This bound is tight in the sense that both the lower and upper bounds are achievable by some distribution.[15] For and as the Bayesian error rate approaches zero, this limit reduces to "not more than twice the Bayesian error rate".
Error rates
There are many results on the error rate of the k nearest neighbour classifiers.[16] The k-nearest neighbour classifier is strongly (that is for any joint distribution on ) consistent provided diverges and converges to zero as .
Let denote the k nearest neighbour classifier based on a training set of size n. Under certain regularity conditions, the excess risk yields the following asymptotic expansion[11]
for some constants and .
The choice offers a trade off between the two terms in the above display, for which the -nearest neighbour error converges to the Bayes error at the optimal (minimax) rate .
When the input data to an algorithm is too large to be processed and it is suspected to be redundant (e.g. the same measurement in both feet and meters) then the input data will be transformed into a reduced representation set of features (also named features vector). Transforming the input data into the set of features is called feature extraction. If the features extracted are carefully chosen it is expected that the features set will extract the relevant information from the input data in order to perform the desired task using this reduced representation instead of the full size input. Feature extraction is performed on raw data prior to applying k-NN algorithm on the transformed data in feature space.
An example of a typical computer vision computation pipeline for face recognition using k-NN including feature extraction and dimension reduction pre-processing steps (usually implemented with OpenCV):
PCA or Fisher LDA projection into feature space, followed by k-NN classification
Dimension reduction
For high-dimensional data (e.g., with number of dimensions more than 10) dimension reduction is usually performed prior to applying the k-NN algorithm in order to avoid the effects of the curse of dimensionality.[17]
The curse of dimensionality in the k-NN context basically means that Euclidean distance is unhelpful in high dimensions because all vectors are almost equidistant to the search query vector (imagine multiple points lying more or less on a circle with the query point at the center; the distance from the query to all data points in the search space is almost the same).
For very-high-dimensional datasets (e.g. when performing a similarity search on live video streams, DNA data or high-dimensional time series) running a fast approximatek-NN search using locality sensitive hashing, "random projections",[19] "sketches"[20] or other high-dimensional similarity search techniques from the VLDB toolbox might be the only feasible option.
Decision boundary
Nearest neighbor rules in effect implicitly compute the decision boundary. It is also possible to compute the decision boundary explicitly, and to do so efficiently, so that the computational complexity is a function of the boundary complexity.[21]
Data reduction
Data reduction is one of the most important problems for work with huge data sets. Usually, only some of the data points are needed for accurate classification. Those data are called the prototypes and can be found as follows:
Select the class-outliers, that is, training data that are classified incorrectly by k-NN (for a given k)
Separate the rest of the data into two sets: (i) the prototypes that are used for the classification decisions and (ii) the absorbed points that can be correctly classified by k-NN using prototypes. The absorbed points can then be removed from the training set.
Selection of class-outliers
A training example surrounded by examples of other classes is called a class outlier. Causes of class outliers include:
random error
insufficient training examples of this class (an isolated example appears instead of a cluster)
missing important features (the classes are separated in other dimensions which we don't know)
too many training examples of other classes (unbalanced classes) that create a "hostile" background for the given small class
Class outliers with k-NN produce noise. They can be detected and separated for future analysis. Given two natural numbers, k>r>0, a training example is called a (k,r)NN class-outlier if its k nearest neighbors include more than r examples of other classes.
Condensed Nearest Neighbor for data reduction
Condensed nearest neighbor (CNN, the Hart algorithm) is an algorithm designed to reduce the data set for k-NN classification.[22] It selects the set of prototypes U from the training data, such that 1NN with U can classify the examples almost as accurately as 1NN does with the whole data set.
Given a training set X, CNN works iteratively:
Scan all elements of X, looking for an element x whose nearest prototype from U has a different label than x.
Remove x from X and add it to U
Repeat the scan until no more prototypes are added to U.
Use U instead of X for classification. The examples that are not prototypes are called "absorbed" points.
It is efficient to scan the training examples in order of decreasing border ratio.[23] The border ratio of a training example x is defined as
a(x) = ‖x'-y‖/‖x-y‖
where ‖x-y‖ is the distance to the closest example y having a different color than x, and ‖x'-y‖ is the distance from y to its closest example x' with the same label as x.
The border ratio is in the interval [0,1] because ‖x'-y‖ never exceeds ‖x-y‖. This ordering gives preference to the borders of the classes for inclusion in the set of prototypes U. A point of a different label than x is called external to x. The calculation of the border ratio is illustrated by the figure on the right. The data points are labeled by colors: the initial point is x and its label is red. External points are blue and green. The closest to x external point is y. The closest to y red point is x' . The border ratio a(x) = ‖x'-y‖ / ‖x-y‖is the attribute of the initial point x.
Below is an illustration of CNN in a series of figures. There are three classes (red, green and blue). Fig. 1: initially there are 60 points in each class. Fig. 2 shows the 1NN classification map: each pixel is classified by 1NN using all the data. Fig. 3 shows the 5NN classification map. White areas correspond to the unclassified regions, where 5NN voting is tied (for example, if there are two green, two red and one blue points among 5 nearest neighbors). Fig. 4 shows the reduced data set. The crosses are the class-outliers selected by the (3,2)NN rule (all the three nearest neighbors of these instances belong to other classes); the squares are the prototypes, and the empty circles are the absorbed points. The left bottom corner shows the numbers of the class-outliers, prototypes and absorbed points for all three classes. The number of prototypes varies from 15% to 20% for different classes in this example. Fig. 5 shows that the 1NN classification map with the prototypes is very similar to that with the initial data set. The figures were produced using the Mirkes applet.[23]
CNN model reduction for k-NN classifiers
Fig. 1. The dataset.
Fig. 2. The 1NN classification map.
Fig. 3. The 5NN classification map.
Fig. 4. The CNN reduced dataset.
Fig. 5. The 1NN classification map based on the CNN extracted prototypes.
In k-NN regression, also known as k-NN smoothing, the k-NN algorithm is used for estimating continuous variables.[citation needed] One such algorithm uses a weighted average of the k nearest neighbors, weighted by the inverse of their distance. This algorithm works as follows:
Compute the Euclidean or Mahalanobis distance from the query example to the labeled examples.
Order the labeled examples by increasing distance.
Find a heuristically optimal number k of nearest neighbors, based on RMSE. This is done using cross validation.
Calculate an inverse distance weighted average with the k-nearest multivariate neighbors.
k-NN outlier
The distance to the kth nearest neighbor can also be seen as a local density estimate and thus is also a popular outlier score in anomaly detection. The larger the distance to the k-NN, the lower the local density, the more likely the query point is an outlier.[24] Although quite simple, this outlier model, along with another classic data mining method, local outlier factor, works quite well also in comparison to more recent and more complex approaches, according to a large scale experimental analysis.[25]
Validation of results
A confusion matrix or "matching matrix" is often used as a tool to validate the accuracy of k-NN classification. More robust statistical methods such as likelihood-ratio test can also be applied.[how?]
^This scheme is a generalization of linear interpolation.
^Hastie, Trevor. (2001). The elements of statistical learning : data mining, inference, and prediction : with 200 full-color illustrations. Tibshirani, Robert., Friedman, J. H. (Jerome H.). New York: Springer. ISBN0-387-95284-5. OCLC46809224.
^Jaskowiak, Pablo A.; Campello, Ricardo J. G. B. (2011). "Comparing Correlation Coefficients as Dissimilarity Measures for Cancer Classification in Gene Expression Data". Brazilian Symposium on Bioinformatics (BSB 2011): 1–8. CiteSeerX10.1.1.208.993.
^Everitt, Brian S.; Landau, Sabine; Leese, Morven; and Stahl, Daniel (2011) "Miscellaneous Clustering Methods", in Cluster Analysis, 5th Edition, John Wiley & Sons, Ltd., Chichester, UK
^Nigsch, Florian; Bender, Andreas; van Buuren, Bernd; Tissen, Jos; Nigsch, Eduard; Mitchell, John B. O. (2006). "Melting point prediction employing k-nearest neighbor algorithms and genetic parameter optimization". Journal of Chemical Information and Modeling. 46 (6): 2412–2422. doi:10.1021/ci060149f. PMID17125183.
^Toussaint, Godfried T. (April 2005). "Geometric proximity graphs for improving nearest neighbor methods in instance-based learning and data mining". International Journal of Computational Geometry and Applications. 15 (2): 101–150. doi:10.1142/S0218195905001622.
^Devroye, L., Gyorfi, L. & Lugosi, G. A Probabilistic Theory of Pattern Recognition. Discrete Appl Math 73, 192–194 (1997).
^Devroye, Luc; Gyorfi, Laszlo; Lugosi, Gabor (1996). A probabilistic theory of pattern recognition. Springer. ISBN978-0-3879-4618-4.
^Bingham, Ella; Mannila, Heikki (2001). "Random projection in dimensionality reduction". Proceedings of the seventh ACM SIGKDD international conference on Knowledge discovery and data mining - KDD '01. pp. 245–250. doi:10.1145/502512.502546. ISBN158113391X. S2CID1854295.
^Ryan, Donna (editor); High Performance Discovery in Time Series, Berlin: Springer, 2004, ISBN0-387-00857-8
^Ramaswamy, Sridhar; Rastogi, Rajeev; Shim, Kyuseok (2000). "Efficient algorithms for mining outliers from large data sets". Proceedings of the 2000 ACM SIGMOD international conference on Management of data - SIGMOD '00. Proceedings of the 2000 ACM SIGMOD international conference on Management of data – SIGMOD '00. pp. 427–438. doi:10.1145/342009.335437. ISBN1-58113-217-4.
^Campos, Guilherme O.; Zimek, Arthur; Sander, Jörg; Campello, Ricardo J. G. B.; Micenková, Barbora; Schubert, Erich; Assent, Ira; Houle, Michael E. (2016). "On the evaluation of unsupervised outlier detection: measures, datasets, and an empirical study". Data Mining and Knowledge Discovery. 30 (4): 891–927. doi:10.1007/s10618-015-0444-8. ISSN1384-5810. S2CID1952214.
هذه المقالة بحاجة لصندوق معلومات. فضلًا ساعد في تحسين هذه المقالة بإضافة صندوق معلومات مخصص إليها. قوس من السحاب يكاد يخفي أشعة الشمس وسرب من الطيور في السماء السحاب القوسي هو نوع منخفض أفقي من أنواع السحاب، ويوجد منه نوعان: سحاب صدفي وسحاب الرول (الأسطوانة المتدحرجة). تتشكل
American politician George Merrick BrooksMember of the U.S. House of Representativesfrom Massachusetts's 7th districtIn officeNovember 2, 1869 – May 13, 1872Preceded byGeorge S. BoutwellSucceeded byConstantine C. EstyChairman of the Concord, Massachusetts Board of SelecmenMember of the Concord, Massachusetts Board of SelecmenIn office1858–1858Member of the Massachusetts Senate Fourth Middlesex DistrictIn office1859–1859Member of the Massachusetts House of Representa...
هذه المقالة يتيمة إذ تصل إليها مقالات أخرى قليلة جدًا. فضلًا، ساعد بإضافة وصلة إليها في مقالات متعلقة بها. (أبريل 2019) تيم روز معلومات شخصية الميلاد 23 سبتمبر 1940 واشنطن الوفاة 24 سبتمبر 2002 (62 سنة) لندن مكان الدفن مقبرة برومتون مواطنة الولايات المتحدة الح...
هذه المقالة يتيمة إذ تصل إليها مقالات أخرى قليلة جدًا. فضلًا، ساعد بإضافة وصلة إليها في مقالات متعلقة بها. (مايو 2014) اضغط هنا للاطلاع على كيفية قراءة التصنيف نقار خشب زيتوني الخلف نقار خشب زيتوني الخلف، Dinopium rafflesii حالة الحفظ أنواع قريبة من خطر الانقراض (IUCN 3.1)[1] المرتبة...
لمعانٍ أخرى، طالع أهل سوس (توضيح). اهل سوس تقسيم إداري البلد المغرب الجهة مراكش آسفي الإقليم آسفي الدائرة عبدة الجماعة القروية لبخاتي المشيخة البخاتي الجنوبية السكان التعداد السكاني 801 نسمة (إحصاء 2004) • عدد الأسر 149 معلومات أخرى التوقيت ت ع م±00:00 (توقيت قياسي)[...
يفتقر محتوى هذه المقالة إلى الاستشهاد بمصادر. فضلاً، ساهم في تطوير هذه المقالة من خلال إضافة مصادر موثوق بها. أي معلومات غير موثقة يمكن التشكيك بها وإزالتها. (ديسمبر 2018) بقعو إحدى قرى محافظة طرطوس السورية، تبعد عن مدينة طرطوس 25 كيلومتر، وعن مدينة الدريكيش 12 كيلومتر، من القرى
Si ce bandeau n'est plus pertinent, retirez-le. Cliquez ici pour en savoir plus. Cet article est orphelin. Moins de trois articles lui sont liés (juillet 2023). Vous pouvez aider en ajoutant des liens vers [[Incident du SOS du mont Asahi]] dans les articles relatifs au sujet. L'incident du SOS du mont Asahi, aussi connu comme l'incident du signal de détresse SOS (SOS遭難事件, Esuōesu sōnan jiken?), est un événement ayant eu lieu en juillet 1989 sur le mont Asahi de la chaîne Daiset...
Libra BintangkuAlbum studio karya Aida MustafaDirilis1967GenrePopDurasi17:43LabelJ&B RecordsProduserJalus SuhermanKronologi Aida Mustafa Aida (1966)Aida1966 Libra Bintangku(1967) Sedjenak Bernjanji Bersama(1967)Sedjenak Bernjanji Bersama1967 Libra Bintangku merupakan album musik kedua karya penyanyi berkebangsaaan Indonesia di era 60-an, Aida Mustafa yang dirilis pada tahun 1967 (meskipun judul albumnya tidak ditunjukkan secara spesifik). Di album ini Aida diiringi oleh Orkes Zainal C...
Darjah Utama Seri Mahkota NegaraDianugerahkan oleh Yang di-Pertuan AgongTipeTanda kehormatanDibentuk16 Agustus 1958Negara MalaysiaMottoBerpegang Tegoh Pada Allah (Melayu kuno)(Berpegang Teguh kepada Allah)Dianugerahkan kepadaRaja Permaisuri Agong dan penguasa di Malaysia, kepala negara asing serta orang-orang terhormat.StatusMasih dianugerahkanPenguasaYang di-Pertuan AgongGelar akhiranD.M.N.StatistikPenganugerahan pertama1958[1]Penganugerahan terakhir2019Jumlah penerima44 penerim...
KabuterimonChampionPenampilan perdanaDigimon Adventure Episode 5[1]PartnerKoushiro Izzumi (Izzy)Evolusi dariTentomonBerevolusi menjadiMegaKabuterimonKongoumon[2] Tingkatan evolusiTentomon Level In-Training Motimon Level Rookie Tentomon Level Champion Kabuterimon Level Ultimate MegaKabuterimon DigiDestined: Koushiro Izzumilbs Kabuterimon merupakan salah satu digimon yang menjadi peran utama dalam Digimon Adventure. Kabuterimon diambil dari kata kabutomushi, sebuah kata dalam ba...
Public health issue of violent acts against women Violence against women in Fiji is recognised to be pervasive, widespread and a serious national issue[1] in the Pacific Island region. Fiji's rates of violence against women are among the very highest in the world.[2] The Fiji Women's Crisis Centre reports that 64% of women who have been in intimate relationships have experienced physical or sexual violence from their partner, including 61% who were physically attacked and 34% ...
TVUIPerkumpulan Televisi Komunitas Fakultas Ilmu Sosial dan Ilmu Politik Universitas IndonesiaPemrogramanBahasaBahasa IndonesiaJaringan televisiIndependen (Komunitas)KepemilikanPemilikUniversitas IndonesiaRiwayatSiaran perdana5 November 2015; 8 tahun lalu (2015-11-05) [1]Siaran terakhirTidak Diketahui [2]Bekas nomor kanal52 UHF (analog)Makna tanda panggilTelevisi Universitas IndonesiaInformasi teknisOtoritas perizinanKementerian Komunikasi dan Informatika RIPranalaSitus w...
1921 conflict after the Treaty of Trianon Uprising in West-HungaryMap of present-day Burgenland in Austria, the area of the short-lived state of Lajtabánság and military conflictsDate28 August - 13 October 1921(1 month, 2 weeks and 1 day)LocationWest-Hungary (Sopron County, Moson County, Vas County)(today part of: Burgenland, Austria and Hungary)Result Hungarian victory and the Sopron ReferendumTerritorialchanges Sopron and its area remained in Hungary.Belligerents Austria Hu...
Australian politician DameDorothy TangneyDBESenator for Western AustraliaIn office21 August 1943 – 30 June 1968 Personal detailsBorn(1907-03-13)13 March 1907North Perth, Western AustraliaDied3 June 1985(1985-06-03) (aged 78)Wembley, Western AustraliaPolitical partyAustralian Labor PartyAlma materUniversity of Western AustraliaOccupationTeacher Dame Dorothy Margaret Tangney DBE (13 March 1907 – 3 June 1985) was an Australian politician. She was a member of the Au...
British television series Princess NikkiStarringNikki GrahameOpening themeHey NikkiCountry of originUnited KingdomOriginal languageEnglishNo. of series1No. of episodes6 (list of episodes)ProductionRunning time30 minutesOriginal releaseNetworkE4Release6 September (2006-09-06) –11 October 2006 (2006-10-11) Princess Nikki is a reality television series that stars former Big Brother 7 housemate Nikki Grahame. It ran on Wednesday nights between 10:00 - 10:30pm on E4 from 6 Septemb...
Антоні Залеський Псевдонім Baronowa X Y ZНародився 2 травня 1858(1858-05-02)[1]Помер 9 серпня 1895(1895-08-09)[1] (37 років)Варшава, Російська імперія[2]Поховання Повонзківський цвинтарКраїна Республіка Польща Російська імперіяДіяльність письменник, журналіст, публіцист...
قرية القعره السفلى - قرية - تقسيم إداري البلد اليمن المحافظة محافظة حجة المديرية مديرية قفل شمر العزلة عزلة بني جل السكان التعداد السكاني 2004 السكان 236 • الذكور 125 • الإناث 111 • عدد الأسر 31 • عدد المساكن 32 معلومات أخرى التوقيت توقيت اليمن (+3 غرين...
Diplomatic summit 2+2 Ministerial DialogueOfficial languagesHindiEnglishJapaneseRussianMembership India United States Japan Australia Russia United Kingdom From left to right: Rajnath Singh, Subrahmanyam Jaishankar, Mike Pompeo, and Mark Esper during the second 2+2 summit on December 18, 2019. The 2+2 Ministerial Dialogue is a diplomatic summit that has been held every year since 2018 initially between the Minister of External Affairs or Foreign Ministe...
1996 single by Manic Street Preachers For other songs of the same name, see Everything Must Go. Everything Must GoSingle by Manic Street Preachersfrom the album Everything Must Go Released22 July 1996 (1996-07-22)[1]Length 3:41 (album version) 3:07 (edit) LabelEpicSongwriter(s)Nicky Wire, James Dean Bradfield, Sean MooreProducer(s)Mike HedgesManic Street Preachers singles chronology A Design for Life (1996) Everything Must Go (1996) Kevin Carter (1996) Everything Must G...