This is the talk page for discussing improvements to the Additive synthesis article. This is not a forum for general discussion of the article's subject.
This article is part of WikiProject Electronics, an attempt to provide a standard approach to writing articles about electronics on Wikipedia. If you would like to participate, you can choose to edit the article attached to this page, or visit the project page, where you can join the project and see a list of open tasks. Leave messages at the project talk pageElectronicsWikipedia:WikiProject ElectronicsTemplate:WikiProject Electronicselectronic
Erm, the unique tone of an instrument is formed more by how those harmonics *change* over time, and by transient bits of noise and non-harmonic frequencies. Additive tries to emulate that by having a different envelope on each individual harmonic. I don't know how best to re-phrase it.
Sorry: I've reworded it a bit further. Hope all is well now. Dysprosia 09:53, 7 Sep 2003 (UTC)
I understand that additive is equivalent to wavetable if partials are harmonic. What if partials aren't harmonic? Which allows me to do this, and which does not? Petwil06:02, 21 October 2006 (UTC)[reply]
wavetable synthesis (not to be confused with basic PCM sample playback cannot do inharmonic partials unless you were to detune the partials from their harmonic frequency by constantly moving the phase of the partial, requiring a lot of wavetable updates. in additive synthesis, the partials (sine waves) are synthesized separately, then added. in their separate synthesis, there is no need that they be at harmonic frequencies. the frequencies of partials in additive synthesis can be whatever is specified. r b-j04:13, 22 October 2006 (UTC)[reply]
The Synclavier was a sampler and programmable harmonic definable wavetable FM synthesizer. It was NOT a real additive synth: you can construct a patch defining 24 fixed partials per voice and apply dynamic enveloping and a very simple FM modulator with envelope, only with the partial timbre upgrade you can specify several harmonic spectrums and fade between them in time. I think the article meant to refer to machines as the Kurzweil K150 or Kawai K5/5000 and remotely the Digital Keyboards Synergy, all them the first generation of additive hardware synths.
The K 150 is a REAL -and a compromise between quantity of oscillators vs. poliphony- additive engine where you can program each partial individually with envelopes (it's a shame that the programming is only possible using an old apple computer, it can't be done from the front panel).
The K5 does the same but is a simplification, being able to control only 4 groups of harmonics and not each one: practice shows that individual control is desirable up to the 16th partial...
The K 5000 is the classic additive synth, but combined with samples: it's quite powerful but clumsy to work compared with software synthesis.
The true about the Synergy:
The Synergy is a user definable PM (as FM) semi algorithmic with additive capabilities, 32 digital oscillators synth. This means that you could use it as a 16 partials two voice polyphony fully additive synth (with limited timbrical results) or the most usual way: complex 8 voice polyphony Yamaha FM style synthesis. You can think of it as a much more flexible algorithm, envelope (for frequency and amplitude for each oscillator) and filter equalization DX7 style synth.
In fact, you can came very close to the original patches using a soft synth as FM7: you cannot do the best patches (as Wendy Carlos's collection) on a DX7 because of the limited envelopes and operator output fixed curves, not to consider the somewhat "metallic" quality of sound that all the DXs have.In comparison, the Synergy is really warm.
That is all and is not small thing.
— Preceding unsigned comment was added by r b-j at 08:17, 13 May 2007 (UTC), and edited by 190.190.31.69 at 20:46, 23 August 2010 (UTC))[reply]
I can't figure out from this article whether additive synthesis involves harmonic partials only, or if inharmonics can be used as well. For example, an early section reads: Additive synthesis ...[combines] waveforms pitched to different harmonics, with a different amplitude envelope on each, along with inharmonic artifacts. Usually, this involves a bank of oscillators tuned to multiples of the base frequency. The term "inharmonic artifacts" implies that they are not deliberate but faults of the technology somehow. The general idea I get here is that additive synthesis is about combining harmonic partials of the fundamental frequency. But further down we get: Additive synthesis can also create non-harmonic sounds if the individual partials are not all having a frequency that is an integer multiple of the same fundamental frequency. Finally, another section says: ...wavetable synthesis is equivalent to additive synthesis in the case that all partials or overtones are harmonic (that is all overtones are at frequencies that are an integer multiple of a fundamental frequency...).
So I'm confused. If I combines a bunch of waves, some of which are not harmonics of the fundamental, is this additive synthesis or not? I'd always assumed it was. Another sentence on this page says: Not all musical sounds have harmonic partials (e.g., bells), but many do. In these cases, an efficient implementation of additive synthesis can be accomplished with wavetable synthesis [instead of additive synthesis]. Yet I've spent time using what I thought was additive synthesis to create bell-like tones, by incorporating various harmonic and inharmonic partials (using sine waves only). It seems like additive synthesis is the right term, since a bunch of waveforms are being "added" together, whether or not they are harmonic. But then again, that's just my uninformed sense. Is there a definitive definition one way or the other? If so, let's edit the page to make that clear. If not, ....let's edit the page to make that clear! Pfly (talk) 09:51, 18 November 2007 (UTC)[reply]
Acoustic instruments and electronic additive synthesizers
This section is a bit of a mess. It delves too deep into the features of a few digital synthesizers and makes quite strongly biased claims about them ("it's quite powerful but clumsy to work", "In comparison, the Synergy is really warm. That is all and is not small thing", "it's a shame that the programming is only possible using an old apple computer"), is ambiguous and jargony at times and generally quite poor in grammar and style. Partially it feels like an advertisement for a synth. Some examples of additive synthesizers would be welcome, but I think this section needs a complete rewrite. Jakce (talk) 11:28, 22 September 2010 (UTC)[reply]
Agreed. I think there needs to be some consensus about whether an acoustic instrument is in any real sense an "additive synthesiser" or just a "Historical precursor." Is it enough to be "a mixture of many sources" to be additive? Does a symphony orchestra count? Is a sum of sinusoids required (as the Theory section suggests)? Is a Hammond tone wheel organ really an additive synthesizer? -- certainly not in the usual sense of a set of time-varying sinusoidal partials. I would be inclined to add a "Historical precursors" section for the pipe organ and hammond. Maybe some of the other stuff belongs there too. Ross bencina (talk) 07:48, 12 January 2012 (UTC)[reply]
I renamed the Implementations section as Timeline of additive synthesizers and moved Wavetable synthesis and Speech synthesis away from there. Organs are now listed as historical precursors. This sectioning should give a better separation between synthesis theory and synthesizer history, so that both can be expanded without compromising the other. I suppose history of advances in theory should be embedded in the theory sections rather than put on the timeline. Olli Niemitalo (talk) 01:30, 13 January 2012 (UTC)[reply]
Additional citations
Why and where does this article need additional citations for verification? What references does it need and how should they be added? Hyacinth (talk) 03:42, 30 December 2011 (UTC)[reply]
It seems slightly strange question. Until December 2010, this article lacked citations at all, and since then, I'm expanding this article and adding most of all citations on the "implementations" section. However, the other sections — the lead section (definition of notion), and resynthesis section (most interesting part) — still lack any citations at all. Your contributions are welcome ! --Clusternote (talk) 09:12, 30 December 2011 (UTC)[reply]
After digging several related references, I felt again that the descriptions on the lead section and re-synthesis section seem to be too naive (not practical), and possibly not based on any reliable sources (except for simple articles for beginners).
Although the descriptions are not incorrect, it seems to be hard to make association with existing reliable researches, and too abstract as foundation for adding extended results researched on several decades ago.
In my opinion, sometime, these sections should be totally re-written.
Okay Cluster, we need to talk about what is meant by "realtime" or "real-time".
I don't think you have the common use or meaning down by "time-variant". "... time-vary transient wave" is time-variant or, if you're more a hardcore techie, nonstationary. What "real-time" means in any context is that the production of whatever is done at the time of consumption, not in advance. It is food that is not pre-cooked or pre-processed but cooked when it is eaten. As far as music synthesis or music processing is concerned, it means that the music is synthesized or processed at the time that it is heard. It means that it was not synthesized in advance and written to a soundfile to be played back later (usually because the computational cost of time in synthesis exceeded the time duration of the sound).
Real-time and time-variant really are different concepts. In music synthesis, one situation I can think of where they are related is in the cranking of a knob (or bending a pitch wheel or mod wheel) during a note. If the synthesis is not real-time and the sound is outputted from a sound file, you cannot do that in playback unless the processing to change the pitch or modulation (whatever the mod is) can be done to the sound playback via real-time post-processing.
I think you need to come up with good references that support your notion that "real-time" means "time-variant". It doesn't. But both concepts have much to do with sound synthesis. 70.109.177.113 (talk) 05:53, 3 January 2012 (UTC)[reply]
Hi, 70.109.177.113. Please create account before discussing, and show your sources for your opinion. I'm not interested on time wasting discussion with unknown person without any reliable sources. --Clusternote (talk) 05:58, 3 January 2012 (UTC)[reply]
Well, that's kinda a copout, Cluster. Consider the merit of the content of the text that appears before you, not whatever disembodied being that placed it there. Why, from what source, did you come up with the notion that in any context, let alone the context of music synthesis, that "real-time" means "time-varying"?
(reset indent)
If you need meaningfull discussion, please show your reliable sources on addtive synthesis at first, then briefly explain your opinion. I can't understand your previous complicated posts.
Note that additive synthesis for dynamic, time-varying waveform generation seems to be historically often called "realtime additive synthesis" in the meaning of both "realtime change of waveform, harmonics or timbre" and "realtime implementation, processing or computation". You can find several examples on Google Scholar search. Or more briefly, Sound on Sound's article on Oct. 1997 show both meaning of usage of the term "real time". --Clusternote (talk) 13:32, 3 January 2012 (UTC)[reply]
Again, I have removed this content change from Clusternote that some might call OR but I would just call an error of category. Clusternote, you are mistaken with your assumption that they didn't mean real-time when they wrote real-time synthesis. What is real-time computing is precisely what is meant in real-time synthesis and your own cites make that connection ("STONE-AGE ADDITIVE"). It's your original contribution to the article that is required to be defended with citations that actually support the addition. 70.109.177.113 (talk) 06:01, 4 January 2012 (UTC)[reply]
Clusternote, I think you're getting confused between what is often implemented at the same time. "additive synthesis for dynamic, time-varying waveform generation" implemented as "realtime additive synthesis" does not mean the two terms are equivalent in meaning. They have distinct meanings as 70.109.177.113 has explained. Ross bencina (talk) 07:53, 12 January 2012 (UTC)[reply]
Dear Ross, thanks to your advice. Probably as you said. Most of our difficulty is lack of reliable source, and moreover, lack of widely accepted common taxonomy of additive synthesis, both required as base of discussion and improvement of article. If you kindly provide several source or list of widely accepted terminologies and taxonomy on this field, several time wasting discussions without any reliable sources by several IP users are drastically reduced, and it should be great help to improve the article. (also discussion on categorization may be found on #Lack of categorization of various .22additive_synthesis.22).
As for "real-time ...": Several years ago, some article also edited by similar authors had denoted "time-varying ..." as "real-time ...", so, I've recognized it as a kind of unique jargon accepted on Wikipedia. However, now, that article is completely clean upped, and that unique expression seems gone. I should had checked the reliability of that expression when I had seen it.
70.109.177.113, please show your reliable sources before editing article. I already show several sources on this issue on this page. You can't understand the situation yet. Almost all citations on article page were added by me. However, you didn't yet show any source supporting your opinion. Please don't revert article until when you can find any sources supporting your opinion. --Clusternote (talk) 06:10, 4 January 2012 (UTC)[reply]
You are the editor adding content without sourcing it. The sources you cite actually disprove the claim you make (that "real-time" in synthesis is not the same as real-time computing. You are totally mistaken and you need to do some studying. Start with the sources you cited above. 70.109.177.113 (talk) 06:21, 4 January 2012 (UTC)[reply]
This user is confusing "realtime sound synthesis"[2] with Real-time computing: On his/her several edits, he/she repeatedly added link to real-time computing for ANS synthesizer[3][4]. However, in truth, ANS synthesizer was implemented using electronic-optical technology, and not using any real-time computing technology. And these incorrect edits are one of the evidences of his/her misunderstandings.
While this user misunderstand repeatedly as above, this user also claimed it as another user's misunderstanding. [5]
Sorry, but I don't think it's appropriate to discriminate against an anonymous IP, even in a talk page. IPs have most of the rights of editing, discussion, consensus, etc. that registered accounts do. There are legitimate reasons to contribute via an IP and it's a core Wikipedia principle to embrace them. Verifiable content is what matters, not who contributes it. Users never need to be trusted, whether they are a named account or an IP. Enough philosophy... back to your argument! :) --Ds13 (talk) 09:14, 3 January 2012 (UTC)[reply]
Thanks for your comment. However, most problematic thing on this issue is, on the article or talk page, other users including this user didn't show any citations or sources, and only discuss on their uncertain memories or original researches. Above discussion essentially lacks sources, and discussion without sources tends to be biased subjectively. We needs reliable sources and responsible discussion. --Clusternote (talk) 10:21, 3 January 2012 (UTC)[reply]
Agree completely. This article needs more citations and less subjectivity. I'm going to make a pass through the article now to neutralize a few things. --Ds13 (talk) 18:28, 3 January 2012 (UTC)[reply]
By the way: If someone understand exact meaning of what this IP user wrote on this thread, please briefly explain it to me.
For me, this IP user's uncertain English couldn't understand well. And in my eyes, his/her uncertain, uncooperative, offensive words seems to be slightly inappropriate on discussion on Wikipedia. --Clusternote (talk) 14:44, 4 January 2012 (UTC)[reply]
Speech synthesis
Hi Clusternote. I understand the point of your recent edit, replacing text I deleted. Here's my perspective: until some reliable sources verify the relevance of the speech synthesis premise (intro sentence of that section?) then I struggle with any and all of its content being there. Will wait for more info. --Ds13 (talk) 02:33, 4 January 2012 (UTC)[reply]
The similarity between "speech synthesis" (imply analysis and re-synthesis) and "additive synthesis" (including resynthesis and analysis needed for it), is just what I already wrote on article. Historically, speech analysis and speech re-synthesis were implemented using extraction of peak frequencies of fomants, and reproduction of peak frequencies using oscillators (in the case of sinewave synthesis, adding sinusoidal waves as in the normal additive synthesis) or filters.
These method didn't directly implement harmonic analysis (and resynthesis based on it), however, the definition of additive synthesis should not be limited on harmonic analysis/resynthesis. A special case of additive synthesis based on harmonic analysis seems to be just called "Spectral modeling synthesis" (or, "Sinusoidal modeling" for the particular case using sinusoidal waves as basis function)[a 1]
P.S. Anyway, thanks for your suggestion. The formal name and categorization of a kind of non-harmonic additive synthesis used on especially speech synthesis, is interesting theme to improve article significantly. I think it requires farther verification. --Clusternote (talk) 08:59, 4 January 2012 (UTC)[reply]
Yesterday I stripped all material that was clearly not related to additive synthesis from the speech synthesis section. There is no need to describe non-additive techniques, formants, CELP etc in this article. On the other hand I can see the relevance of the content that I've left. Possibly it would be better to replace with a related techniques section that includes the current content of the speech synthesis section. After all, additive synthesis is related to many other areas, not just speech synthesis research. --Ross bencina (talk) 05:45, 14 January 2012 (UTC)[reply]
The latest "improvement" reverted.
I will certainly not be opposed to putting in some good references for the theory in the article, but I am just as certainly opposed to putting in incorrect mathematics. Clusternote, you made a common error. In the expression:
the parameter is not the (angular) frequency because it is what multiplies . It is the the frequency because it's the time rate of change of the internal phase. You take whatever is inside the function and you differentiate it to get . You do not take whatever is inside the function and divide it by to get the frequency. It might make no difference when is constant, but it makes a big difference when the instantaneous frequency varies in time. 70.109.180.242 (talk) 18:04, 7 January 2012 (UTC)[reply]
I meant to add that some of the changes in prose that you made might very well be good. But removing the summations (which is the discrete-time way of integrating) and replacing by the simpler multiplication was incorrect. (I think it was Einstein that said "A system should be as simple as possible, but no simpler."). 70.109.180.242 (talk) 18:10, 7 January 2012 (UTC)[reply]
Stop vandalism! This article not contains expression or formula modification you said. Your story is not related to this article at all.--Clusternote (talk) 20:08, 7 January 2012 (UTC)[reply]
I've also made this error in the past, so I can understand it got slipped in in error. But we should keep the brackets that Clusternote added around the contents of the summations. Please assume good faith, both sides; I don't see anything to the contrary. Olli Niemitalo (talk) 16:45, 10 January 2012 (UTC)[reply]
Can this please be fixed (the Fourier series equation missing brackets) as it is quite clearly wrong, in fact the same, correct equation appears in the wiki article on Fourier series. You can also use for clarity and to match the other article.
Stop reverting, 71.169.179.65. This article is basically already added reliable sources, except for the description tagged with [Citation needed].
I don't want to own this article. I only add reliable sources and correct several problems. You have no right to revert article without any reliable sources. --Clusternote (talk) 17:48, 9 January 2012 (UTC)[reply]
Practice what you preach.
Even though I work for a living, I am willing to take a little time to discuss. But I will not let you (as best as I can) take over the article and turn it into your sole POV. 71.169.179.65 (talk) 17:57, 9 January 2012 (UTC)[reply]
Your opinion & POV always lacks any reliable sources. I always show/add reliable sources. You should not revert article based on your POV.
Your repeatedly reverts without preceding discussion nor reliable sources, are equivalent to vandalism.
Practice what you preach.
If you want to keep your equations on Additive synthesis#Theory section, you should show reliable source on your very unique equations. Especially, equations on Additive synthesis#Non-harmonic signal seems to be inappropriate on Wikipedia, because treat of time-varying frequency factor (on these equation, these are time-averaged) lacks generality and accuracy. Your private, unpublished algorithm is inappropriate on the Wikipedia article. --Clusternote (talk) 18:32, 9 January 2012 (UTC)[reply]
They're not my equations, but they are correct. I cited them with two articles from Julius Smith, and I told you why your equations (where you simply multiply the time-variant frequency with t ) are wrong. You dismissed that with no understanding of what you are talking about. 71.169.179.65 (talk) 18:38, 9 January 2012 (UTC)[reply]
On simple substitution k f0 → fk[n], his real substitute (2 π k f0 / Fs) → (2 π / Fs)∑ni=1fk[i]
Note: On the discussion of corresponding section, the context requires substitution of constantk f0 with inharmonic version (on here, "inharmonic" means "arbitrary") fk, or time-varying inharmonic version fk[i]
He declared substitution of constantk f0 with time-varying inharmonic version fk[i] .
And his real substitution is ∑ni=1fk[i] / n imply time-average of time-varying inharmonic version. (if he intent to extend equations towards "time-varying inharmonic" version, this average operation spoil the time-varying nature)
His word and real substitution are contradicted, and also contracted with context.
If YOU think it is correct and no farther explanation is needed, probably YOU are just the person who posted these equations in 2007.
I took a quick look at the diff and that IP (207.something) is not me. They look similar but not identical to the current ones. And they are obviously correct. If you can do math you would know this. They are precisely the discrete-time equivalent to what Julius Smith has published. Would you like me to give you another reason why your equations (I can get the diff, but we both know you tried to add them in the last couple of days) are incorrect? Consider what happens a second or two after the note begins and n = 44100 or 88200 or some big number like that. Then consider what happens to the phase when the instantaneous frequency changes a little. 71.169.179.65 (talk) 18:56, 9 January 2012 (UTC)[reply]
We don't need time-average here.
Red herring. The only sense of "time averaging" is that of the instantaneous frequency over one sample period. It doesn't change anything.
Also I checked source cited on, however, what your saying is not found anywhere. Verification failed, therefore, these expression should be deleted. --Clusternote (talk) 19:03, 9 January 2012 (UTC)[reply]
No you didn't. It's clearly referenced at [6] and [7]. And the latter reference is discrete-time. But the former reference is continuous-time and it translates directly to the discrete-time version when you substitute t = nT = n(1/Fs).
As a result, on these pages, at least no clear explanations about time-averaging operation of frequency. If you think "it's clearly referenced", just quote it here.
Your reply is falsehood. "Time-average ... over one sample period" means ∑1i=1 fk[i] = fk[1] . On such case, time-averaging is not needed. --Clusternote (talk) 19:37, 9 January 2012 (UTC)[reply]
Falsehood by User:71.169.179.65
User:71.169.179.65 posted false statement on this talk page:
I've fully protected the article for three days. Please continue the discussion on the talk page and if you can't come to an agreement, ask for advice from an editor experienced in the area of mathematics. —Tom Morris (talk) 19:06, 9 January 2012 (UTC)[reply]
Can I step in and make a few comments? Firstly, "real-time sound synthesis" is definitely an application of "real-time computing". In the particular case of sound synthesis, the time limitation is usually the sample interval, or a block of them. Given that samples go out at (for example) 48KHz, I have only a limited amount of time to calculate the next sample - about 21us in this case. This is the "real-time constraint" or "operational deadline" that the Realtime computing article talks about. If the sample isn't ready by then, you get a glitch. I don't know exactly what caused the row about that, but it's simple to clear up.
The equations given seem like pretty standard Fourier stuff. The first one given in each case expresses each harmonic as a sum of a sine and cosine, whereas the second expresses it with a variable phase. These are equivalent. There are probably a lot of very-slightly-different-but-equivalent ways of expressing these, but again, it shouldn't be causing this kind of conflict. It's simple to look up.
The only person demanding credentials was Cluster, when he insisted that I create an account and that he wouldn't "waste time" with me without it. I have only been demanding competence and accuracy. There is a substantive difference when Cluster tried to "simplify" the equations for non-harmonic additive synthesis, where the instantaneous frequency of the partials may be time-varying, and it is a difference where we both have insisted the other is clearly wrong. There is no way we can both be correct about that. I have spelled it out to Cluster (and everyone else to see), both in what the general idea is which he dismissed as "Your story is not related to this article at all." (just like "real-time additive synthesis is not related to real-time computing", boy is this guy ignorant) and I also spelled out how his "simplified" equations are in error.
If you want, I can try to elucidate, but I really am tired of talking. I just don't want this article to decline further, which is why I took some effort to challenge an ignorant editor who fancies himself an expert when there were clear errors. I don't mind other people editing, just don't screw it up. And if you don't know something, leave it alone. 71.169.179.65 (talk) 06:57, 10 January 2012 (UTC)[reply]
Thanks for your help. At now, we discuss about equations on sub-section "Non-harmonic signal". There are, 2 versions, basically:
I believe the previous version (edit 470480517) to be mathematically correct, and that the representation of the phase term used in this version is supported by references [2] and [3].
The equations in the new simple version (edit 470480676) are incorrect, since they the a form of equation for a sine wave (which is valid only if the frequency is constant in time) to describe a sine wave with a time-varying frequency. I do not believe that references [2] and [3] support this form of the equations.
This is not to say that the previous version (edit 470480517) is perfect; I think that it could profitably be shortened (the equivalent forms of the equation could perhaps be stated without showing exactly how each is derived from the others). Also, given the section title of "non-harmonic signal", the distinction between constant non-integer multiples of a constant fundamental frequency (partials as opposed to harmonics, in the sense of [[8]]) and signals with time-varying frequency could be made clearer. Mathematically though, the previous version (edit 470480517) is correct and 'new, simple version' is incorrect. Credentials: (since Electricdruid put his down...) Ph.D in mathematics, currently employed as a researcher in mathematics. Chrisjohnson (talk) 22:39, 9 January 2012 (UTC)[reply]
I have had a look down the article and it looks fairly bad.
The first chunk of mathematics is presumably OK, but a bit pompous given knowledge of what a Fourier series is. The second chunk may not be needed, or not there anyway.
I'd like to re-order the whole thing, starting with the theory of the "harmonic signal", then something like the bit about "Additive resynthesis"; and then the examples that are appropriate for the harmonic signal. After that the quasi-periodic signals; and speech synthesis last of all.
I don't know how easy it would be to reference everything in the current article properly. I do know that addressing writing issues is much easier once there is a clear logical flow. Charles Matthews (talk) 21:29, 9 January 2012 (UTC)[reply]
Thank you to the experts trying to clean up and verify the mathematics here. I have a question... since many readers may not be inclined or able to work through the math necessary to verify equations... isn't sourcing those equations from an existing reliable source the best idea? --Ds13 (talk) 05:34, 10 January 2012 (UTC)[reply]
Ds, User:Clusternote is not an expert. He doesn't know shit and he has been crapping up the article immensely.
The language in the discussion of the mathematical basis of additive synthesis may sound "pompous", I didn't write it, nor did I write the original equations, but I might have touched them up in the past. I don't remember. I think the pompous language was really more condescending than pompous. I believe it may have been so because it may have originally been written by computer musicians who may have marveled more at the math regarding Fourier series than mathematicians, scientists, and engineers do. This is also why (in my estimation) there was this lofty reference to "Fourier's theorem" or such, that I yesterday simply changed to Fourier series in the lede.
Now, given that the consumers of this article will be both the hard-core techies that might know a lot about Fourier analysis and little about music and the computer musicians that would have that strength in expertise turned around, I would say that the article would do well to clearly connect the Fourier series concept first to harmonic tones (these would be quasi-periodic, but not in the strict sense that mathematicians mean, but closer to the almost periodic functions, but the use of the term has been used in the audio signal processing and computer music lit for years) and then to generalize to non-harmonic additive synthesis. Now to do that rigorously, we would have to first start out with perfectly periodic real signals (at least while the synthesizer key is depressed) that would have cosine and sine terms in continuous-time with frequencies going up to infinity. Then we would generalize a little and make the tone quasi-periodic by allowing the ak and bk coefficients to be slowly changing functions of time (like envelopes in a synthesizer). Then we would have to limit the top frequency (and the upper limit of the Fourier series summation) so that the signal would be bandlimited. Then we could sample it by substituting t = nT = n(1/Fs). Those four beginning steps were obviously left out from the article, even from its very beginning. Charles, do you see a good reason to include those steps? I didn't, otherwise I would have added them some time ago. But if you do and want to really make this look like a tutorial, I'm okay with it. My major concern was to go from a concept that looks similar to Fourier series and get to the equations that define exactly how we program the sample processing code that performs additive synthesis and the following does that. What words and citations one dresses that up with is less of a concern for me as long as it is both accurate and readable.
So, I'm plopping this down as a starting point and ask that you, Charles, quickly delete or correct the incorrect math that Clusternote has ignorantly replaced it with. It may have started out as WP:good faith but, because he refuses to recognize and admit that he just doesn't get the math, insisting that he canonize his ignorance in the article (and calling any alternative "false" or "incorrect" or "vandalism") is no longer good faith. I will not deal with him anymore because he is not a "straight shooter". He is fundamentally dishonest, and the only thing to do now is simply call him on it (and then ignore him).
BTW, I'm confident of your mathematical chops, Charles, if there are questions you have about the practice of additive synthesis, what or how we coders deal with it, feel free to ask any. 71.169.179.65 (talk) 06:20, 10 January 2012 (UTC)[reply]
I forgot to mention a common notational convention in digital signal processing that you might not be aware of, Charles. For discrete-time signals, rather than put the discrete-time index, n, in a subscript (we might have a vector of signals as we do here and each element might be subscripted already), we show n as an argument, just like the continuous-time t, but we put the discrete-time argument into square brackets to indicate that there should only be integer values going in there. So it's x(t) and x[n] where t = nT. — Preceding unsigned comment added by 71.169.179.65 (talk) 06:39, 10 January 2012 (UTC)[reply]
So, what would you add or subtract from the following mathematical development?:
I would certainly rewrite it to be somewhat more accessible to the "general reader" (point 1); the usual way is to use a phrases like "in other words, a periodic signal is written as a superposition sum of sine waves ...". Charles Matthews (talk) 08:29, 10 January 2012 (UTC)[reply]
So would you do it? Presently the section Non-harmonic signals as it stands is fully incorrect, for the reasons that I and two other editors have pointed out. If time is a problem, can you at least delete that section so that Wikipedia does not look so stupid in the meantime? 71.169.179.65 (talk) 18:17, 10 January 2012 (UTC)[reply]
Basic mathematical foundation of additive synthesis, from Fourier series to non-harmonic synthesis
Harmonic tone
or
where
y[n] is the output sample at discrete time n,
ak[n] = rk[n] cos(φk[n])
bk[n] = rk[n] sin(φk[n])
rk[n] = (ak[n]2 + bk[n]2)1/2 is the amplitudeenvelope of the k-th harmonic at discrete time n,
φk[n] is the instantaneous phase function of the k-th harmonic at discrete time n,
K is the number of harmonics; K < floor(Fs/(2 f0)),
Kf0 is the frequency of the highest harmonic and is below the Nyquist frequency, Fs/2.
The DC term is normally not used in audio synthesis, so the a0[n] term can be removed. Introducing time varying coefficients rk[n] allows for the dynamic use of envelopes to modulate oscillators creating a "quasi-periodic" waveform (one that is periodic over the short term but changes its waveform shape over the longer term).
Non-harmonic tone
Additive synthesis can also create non-harmonic sounds (which appear to be non-periodic waveforms, within the time-frame given by the fundamental frequency) if the individual harmonics do not all have a frequency that is an integer multiple of the fundamental frequency.
By replacing the k-th harmonic frequency, k f0, with a time-varying and general (not necessarily harmonic) frequency, fk[n], (the instantaneous frequency of the k-th partial at the time of sample n) the definition (removing the DC term) of the synthesized output would be:
or
where
.
If fk[n] = k f0, with constant f0, all partials are harmonic, the synthesized waveform is quasi-periodic, and the more general equations above reduce to the simpler equations at the top. For each non-harmonic partial, the phase term φk[n] can be absorbed into the instantaneous frequency term, fk[n] by the substitution:
If that substitution is made, all of the φk[n] phase terms can be set to zero with no loss of generality (retaining the initial phase value at y[0]) and the expressions of non-harmonic additive synthesis can be simplified to
.
If this constant phase term (at time n = 0) is expressed as
the general expression of additive synthesis can be further simplified:
where
and
The two citations in the first para of "Inharmonic Partials" do not link to the citaions that they name. It is unclear to me whether it is the name or the link that is incorrect.
P.S. In my views, without clear explanation of these on article, above complicated equations (which use time-averaging operation to erase phase term) is not rationalized. --Clusternote (talk) 11:01, 10 January 2012 (UTC)[reply]
Clusternote: The sum is not a time-averaging operation -- it's a time-integrating operation. The disputed term is: . This is the sum from (the start time) to (the current time) of (the angular frequency at each sample) divided by the sampling rate . This sum is the discrete version of the integral term in the third equation in reference 3 [[9]] which reads . We can prove that the discrete form of the equation follows from the continuous form in reference 3 by defining the discrete angular frequency at sample , , to be the mean of the continuous angular frequency over one sample, i.e. the integral with respect to time of the continuous angular frequency over the duration of sample , divided by the length of a sample. Chrisjohnson (talk) 12:01, 10 January 2012 (UTC)[reply]
Thank you to kindly explanation. Probably I should check the deliberation of equations written on that source. I'll ask you farther details on the later. --Clusternote (talk) 12:51, 10 January 2012 (UTC)[reply]
I am sorry to trouble you. In my eyes, on the section "Non-harmonic signal", at least short explanations on the following points are necessary to clarify the definition. Without these, the intention of definition is hard to understand.
The phase term is discarded (or redefined as following), and when needed, it is redefined as the integral of the instantaneous frequency [1]
(Note: the roles of and seems inverted on 2nd reference,[2] (wow!) so corrected)
[incorrect]: On the discrete definition, discrete form of above integral form is used, instead of original time-varying form, because ...
[ERRTA]: On the discrete definition, discrete form of above integral form, + (or, + (?)) is used, instead of original time-varying form, because ...
Question: Why discrete form of integral form is used ?
It is because the given discrete formulas describe an implementation. With a typical sampling frequency of 44100 Hz, the synthesis parameters (frequency and amplitude) change so slow that they can be approximated as constant over each sampling period, with little loss in audio quality. Integrating over the piece-wise constant function can be written as a sum. Olli Niemitalo (talk) 10:02, 13 January 2012 (UTC)[reply]
Thanks, but, realy ? On that case, isn't it something like a (where sampling interval ) ?! ... I should personaly verify it, and also think more about it ... --Clusternote (talk) 03:39, 14 January 2012 (UTC)[reply]
P.S. Possibly, are you meaning "efficiency of code using constant (pre-calculated finite integration) instead of array inside summation" ? If so, I understand it, apart from integration range. (I didn't expect "efficiency of implementation" instead of "theoretical possibility of mathematical model" on the "Theory" section. best regards, --Clusternote (talk) 05:37, 14 January 2012 (UTC)[reply]
There's both and efficiency aspect and a simplicity aspect. Synth programmers do not necessarily even think of what happens between the sample points for the slowly-varying control signals, but operate directly on the discrete signals. But you are right in that the introductory theory section should not be concerned with such matters. That's why it would be better to write the continuous formulas there and have the discrete ones in an implementation section. Olli Niemitalo (talk) 07:40, 14 January 2012 (UTC)[reply]
I erased the first footnote (going to have a look at the second one next) as it dealt with instantaneous phase which is an analysis concept not needed in synthesis. The phase of cosine or sine is its argument, so the route via analytic signals is unnecessary. There is a concept in the current theory called phase offset (phase term). Depending on how the synthesis equation is written, it can be constant or time-dependent. Discretization is what has to do with efficiency, not instantaneous phase which is a concept that is not required here. Olli Niemitalo (talk) 11:01, 14 January 2012 (UTC)[reply]
Sorry, but I've slightly changed my opinion on it. Several mentions on instantaneous phase seems to be still needed on supplemental note using continuous form for inharmonic section. Without it, on supplemental note, we can't simply explain the intention of equations. --Clusternote (talk) 03:25, 16 January 2012 (UTC)[reply]
Thanks for your comment. Several miss are correctable as tail of this message.
I think latest version(except for an expression shown in the bold italic letter on the below)(with draw proposal inside paresis due to my mistake (must be instantaneous phase×constant) + additional clarification done on article (15 Jan. 09:45(UST))) is better for engineer preferring discrete form, however, also the concepts behind them should be described for readers more familiar with continuous form. And also possibly other readers will be convinced when they found equivalent "generic expression" were led by both descriptions (article body, and footnote). --Clusternote (talk) 15:18, 14 January 2012 (UTC)[reply]
For readers not familiar with above notations, descriptions using these are possibly not easy to understand at a glance. The following is a try to provide supplemental explanations for these readers.
In continuous form, a wave (correspond to a partial on the 2nd equation) is expressed as
Why the absorption of phase term to instantaneous frequency is needed ? (probably to be consistent with above phase redefinition ? It is also Question)
P.S. Moreover, this specific definition clearlyseems eliminate more generic explanation like a "summation of sinusoidal waves", or "explicitly adding sinusoidal overtones together" written on lead section of article.
According to Smith III 2011, this specific definition was called and defined as:
Additive Synthesis as Early Sinusoidal Modeling:[2]
The former specific definition seems slightly too strict for this article, especially, without enough rationalization of phase redefinition (it cause integration) mentioned on previous my post. I'm glad if I can also read your opinion on it. sincerely, --Clusternote (talk) 06:48, 12 January 2012 (UTC)[reply]
Personally, I would be happy to have the equations written in a continuous-time form, as they are in Smith III 2011. The continuous-time form is probably more natural for most people that are not DSP specialists (but otoh the discrete-time form is going to be of more use to anyone wanting to code up the algorithm on a computer). As for absorbing the phase term, I think a bit too much is made of this point at the moment, and the emphasis should remain on the key process of additive synthesis. On that note, I would favour getting rid of the form in the time-varying/inharmonic section and leaving just the form. It seems to me that the reason to use the form is largely to emphasise the link with a Fourier series / FFT, which is really relevant only for harmonic additive synthesis. Chrisjohnson (talk) 18:08, 12 January 2012 (UTC)[reply]
I agree with all of Chrisjohnson's points. The discrete equations could still be left in concise form in an Implementation section closer to the end of the article. Olli Niemitalo (talk) 18:58, 12 January 2012 (UTC)[reply]
I agree, too. I just don't want them changed to non-correct equations. Being an IP, I cannot import in an image, but could someone draw out a simple diagram with a big summation in the middle and a bunch of oscillators, with frequency and amplitude control, going into that summer? Maybe we could do the theory at the beginning with continuous-time equations, and then in an implementation section, show how the theoretical continuous-time equations get sampled and turned into the simplest and most general equations that currently live at the bottom of the Theory section. 70.109.183.99 (talk) 21:18, 12 January 2012 (UTC)[reply]
Image:SSadditiveblock.png I've transferred File:SSadditiveblock.png from wikibooks to Wikimedia/Commons, with the help of CommonsHelper 2 on toolserver. (At the moment, it temporally marked as "insufficient license" due to my small miss-translation, however, it should be kept after reviewing, because already properly fixed. (according to w:commons:Commons:Copyright_tags#General, I should select {{PD-user-w}} when I transferred PD-user image from wikibooks (other project)) --Clusternote (talk) 05:20, 14 January 2012 (UTC)[reply]
I'm glad for your comment, Chris. I also think discrete expressions should be kept, and additionally, supplemental explanations mentioned on above my post are also needed to clarify the intention of definition for readers.
Since 2007, these equations had been almost unverifiable due to lack of sources. A several days ago, by two additional citations[2][1], the intention of equations seems partially clarified, at least in my eyes. And now, it should be more clearly explained as a definition with several rationality, instead of as inviolable theory. Also, whether this specific definition is only available one or not, seems need to verify.
(Possibly a few peoples on this field claim their theory or definitions don't need any verifiability nor farther supplemental explanations, however, it may not meet Wikipedia standard)
Yes please, if you have a rewrite of the harmonic & inharmonic theory section or parts of it, please post. Currently it is hard to see the proposal from the comments. Olli Niemitalo (talk) 10:59, 13 January 2012 (UTC)[reply]
P.S. As for the questions on above my proposal: superficial short answer may be "by definition" and "to get 'generic expression' often used/required on somewhere" (although these superficial answers don't meets my essential question — rationality of definition and description). And essential answers may be, for example, "efficiency of implementations on DSP hardware/code", "understandability" (without interpretation of continuous formula), or "ease of development" (discrete formula is convenient as basis of programming) as you said. I expect several of them can be verified on reliable sources.
On the definition of mathematical model, too long formula modification without enough explanation ("where are we and where are going?"), tends to cause unnecessary difficulty and misleading. In this case, precondition of definition, intention of definition and modification, and merits of generic expression, should be supplementary explained on the theory section, in my opinion. --Clusternote (talk) 10:24, 13 January 2012 (UTC)[reply]
Opinion of other users
Because, with the exception of organ pipes or or Hammond B3 tonewheels or a collection of analog oscillators (in the 70s, I did a simple 3-harmonic additive synthesizer with a patchboard Moog synthesizer), when additive synthesis is normally done nowadaze, it's done with computers or DSP chips or related chips (like an ASIC or FPGA). These digital synthesizers (or synthesizer software) compute and output discrete-time samples not over continuous time. So the mathematics of what is really being done is not going to be:
but you substitute which is what we do when we sample a continuous-time signal. Now, can you show us that when you make that substitution, you will get an expression that looks like:
?
There will be a slight change in definition for rk and fk, but the form is that.
Now, there is a reason for you to show us that you can do that. You should show us that you know the least bit about mathematics so that we don't think we're wasting our time trying to inform you of an arcane technical concept that you cannot understand (at least understand in a ready amount of time). You cannot demand that your declaration (that previous math is wrong) be respected just because you don't understand it. And you cannot expect to change the article from correct equations that you cannot understand to incorrect equations that you do understand. That's a ridiculous demand. It's a ridiculous fallacy to assert an argument that "whatever I can't understand must be incorrect." 70.109.183.99 (talk) 05:09, 12 January 2012 (UTC)[reply]
Maybe we'll answer this latter question once you've demonstrated that you're worthy (by showing us you can make that substitution and get the discrete-time result). Our time is too valuable to waste. If you cannot understand even the simplest fundamental concepts, what point is it in trying to teach you the somewhat less simple concepts. And your understanding is not the criterion for inclusion. You do not own the article. 70.109.183.99 (talk) 05:09, 12 January 2012 (UTC)[reply]
Personal attack from IP user (moved from Mathematician section)
Thanks for not doing this to the article, Clusternote. It's good that you want to check out and understand the mathematics. It's calculus-level math and it shouldn't be so hard to personally think about and confirm the math. But if you cannot do that yourself, do not conclude that you are the authority of the facts and change it in the article or crap up the article with stikeouts and the like. The main namespaceAdditive synthesis page is not a sandbox.
Overview of Additive Re-synthesis based on (Horner & Beauchamp 1995), (Yee-King 2007, p. 19) [I think this diagram specialises too much on analysis/resynthesis to be described as an 'overview' of additive synthesis. For example, Yee-King 2007, p. 18, mentions using ADSR envelopes rather than analysis to generate the amplitude of the partials. Perhaps the diagram can be expanded to include several possible sources of control data? Chrisjohnson (talk) 18:20, 10 January 2012 (UTC) ][I think this diagram would work well floated right in the Analysis/Synthesis section. With new title "Overview of Additive Analysis/Resynthesis"Ross bencina (talk) 16:25, 13 January 2012 (UTC)][reply]Overview of
R. J. McAulay and T. F. Quatieri (Aug 1986), "Speech analysis/synthesis based on a sinusoidal representation", IEEE Transactions on Acoustics, Speech, Signal Processing, ASSP-34: 744–754
The "Analysis" phase is not an essential part of additive synthesis. It might be an essential part of additive *re*synthesis, but not additive synthesis. Almost any synthesis technique could be used for resynthesis of some existing sound (with varying degrees of success, obviously!) and the analysis stage would be crucial in that process. But the analysis is not part of the synthesis. Additive synthesis can be used for abstract sound synthesis like any other technique, in which case no analysis is required. Leave it out, in my view. Electricdruid (talk) 14:47, 10 January 2012 (UTC)[reply]
Thank you to rapid comment. I know several peoples often oversimplify the definition, however, I can't found enough reliable sources supporting your opinion disregarding analysis phase. Where is your reliable sources ?
As you can see, there's nothing about analysis at all. Columbia Uni is a reliable source, I'd have thought. Where and how you generate control data for your additive synthesis is entirely up to you - most of the instruments you've included as examples allow various sources for the data, and many don't include any analysis features.
I agree with Electricdruid that the analysis phase is not an essential part of additive synthesis. A reliable source for this is 'The Computer Music Tutorial' by Curtis Roads (MIT Press 1996). Many of the examples of additive synthesis on the page, such as the Hammond organ, clearly have no analysis phase -- and surely these should be classified as additive synthesis? Of course, analysis of the sort described above is sometimes used to generate envelopes for each partial for subsequent additive synthesis, but the references Horner & Beauchamp and Yee-King do not support the view that it is the only way to generate such control data. (An example: the THX sound is a frequently-heard instance of inharmonic time-varying additive synthesis being used with control data generated entirely algorithmically, rather than through analysis of existing sounds). Chrisjohnson (talk) 15:34, 10 January 2012 (UTC)[reply]
Thanks for your comment and reference. That difference may be wider meaning (as Analysis/Resynthesis-style synthesis) and narrower meaning (concentrate to the form of "Fourier series" originally used on re-synthesis phase). Narrower usage is often used on consumer models, however, researches seems to be carried out regarding wider viewpoint. For example, (Smith III & Serra 2005, Additive Synthesis) defines "additive synthesis" with referring "Speech analysis/synthesis" paper (McAulay & Quatieri, 1986). --Clusternote (talk) 15:58, 10 January 2012 (UTC)[reply]
I take the view that the 'wider meaning' of additive synthesis is the summation of sinusoidal waves (a Fourier series or otherwise) with amplitudes and frequencies produced by any method, and that it is a 'narrower view' to limit attention to the situation where the amplitudes and frequencies of the partials are generated by some form of analysis of other sounds. In an article titled Additive Synthesis we should not restrict attention only to systems where additive synthesis is combined with analysis. It is misleading to regard additive synthesis as being something "originally used on re-synthesis phase" [of an analysis/re-synthesis system], since additive synthesis without any analysis pre-dates any form of analysis/re-synthesis by a considerable period (e.g. the Telharmonium of 1897). (By the way, Smith & Serra cite McAulay and Quatieri only as a reference for cubic polynomial phase interpolation -- I'm not sure this is a relevant source?). Chrisjohnson (talk) 16:39, 10 January 2012 (UTC)[reply]
My selection of words "wider" and "narrower" are possibly opposition.
On (McAulay and Quatieri 1986), your point seems reasonable. What I had to point out was, at least, several researchers of that field seems to be interested in Analysis/Resynthesis and Speech synthesis (= wider viewpoint), for example, title of Smith III & Serra 2005 "PARSHL: An Analysis/Synthesis Program for Non-Harmonic Sounds Based on a Sinusoidal Representation" show it. And now, I think farther verification is needed on this issue including Analysis/Resynthesis on earlier researches of additive synthesis, and also (possibly) interaction with research of Speech Synthesis, etc. --Clusternote (talk) 19:05, 11 January 2012 (UTC)[reply]
I think the subject of analysis merits inclusion in the article, but it should be made clear that it is not a part of all additive synthesis methods. Also a note on the use of the word "decomposition": In this context it is synonymous to "analysis", not to "modification" (a use not supported by the given references). The idea is that the signal is thought to be composed of (typically sinusoidal) components, and can thus be decomposed back into those components. See for example the abstract of Simmons 2002. Olli Niemitalo (talk) 16:14, 10 January 2012 (UTC)[reply]
Thanks for your comment. I think so. For the term "Decomposition", it may be probably my mistake, or non-standard use of terminology on somewhere. I temporally missed reference to it. When I found the URL, of course, I want to verify it. --Clusternote (talk) 17:14, 10 January 2012 (UTC)[reply]
I've cleaned up the Additive analysis/re-synthesis section based on the above discussion. Can we remove the cleanup tag now? Perhaps the illustration above could be included as "A schematic illustration of the additive analysis/resynthesis process".
Ross bencina (talk) 12:22, 13 January 2012 (UTC)[reply]
Clusternote, Chris, others: I think there is no such thing as "resynthesis" aside from if there is a preceding *analysis*. Any other form of control can only lead to "synthesis" (controlled synthesis, if you will). For this reason I think that while Clusternote's "new simplified" diagram is good and correct as a "overview of timbral specification techniques for additive synthesis" I don't think it belongs with "resynthesis" or "Analysis/resynthesis". Perhaps we should create a new "Control Strategies" section. I don't know when the current section was originally titled "Additive resynthesis" but I have now renamed it "Additive analysis/resynthesis" since that made the most sense based on incorporating all of the discussion in this section. Of course there is still a question about whether analysis/resynthesis should be mentioned at all. I think it is a significant enough area to mention -- so long as it isn't presented as the whole story, which it isn't. Ross bencina (talk) 05:05, 14 January 2012 (UTC)[reply]
Lack of categorization of various "additive synthesis"
For example, according to Horner & Beauchamp 1995, they used "wavetables" and "static additive synthesis" to speed up the computation, and their "time-varying additive synthesis" was replaced by interpolation between wavetables.
Real-time control or not
Real-time control ready (possibly Real-time additive synthesis)
No real-time control (concerning additive synthesis)
TheseSimilar categorizations and several explanations are needed on article on appropriate section. , beforehand implicitly define additive synthesis without enough reliable sources. --Clusternote (talk) 14:07, 10 January 2012 (UTC) [updated]--Clusternote (talk) 18:08, 11 January 2012 (UTC)[reply]
Comments on the proposed classification: I think the distinctions of harmonic/inharmonic, steady-state/time-varying and mathematically-generated/human-generated/analysis-generated control data for the partials are important, and need to be clear from the article. However, I think the article initially needs to focus on what is common to these different approaches (the summation of sinusoidal waves), before distinguishing the various types of additive synthesis. I don't think the 'real-time' classification is required - the realtime issue is only relevant in that the high computational cost of generating and summing many sinusoidal partials on a digital computer/synth motivated the development of a number of related techniques that are less computationally expensive. Chrisjohnson (talk) 19:08, 10 January 2012 (UTC)[reply]
I don't think categorization is that important, especially if there is no common agreement on the names of the categories. As a workaround, one could section the article into the problems/questions that would have lead to the categorizations, and discuss the alternative approaches there (not necessarily naming the categories): Frequencies of the partials, Source of control signals or something to that effect. Real-time control is something that can be mentioned, if needed, in plain speak for each synthesizer model listed. It's more like a practical point rather than anything fundamental in the theory: Will the sound change as I turn the knob? Olli Niemitalo (talk) 20:30, 10 January 2012 (UTC)[reply]
I'm glad for your comment. Really, lack of common agreement of categorization on additive synthesis seems important issue. Rough notion of each categorization seems to be almost shared, however, detail and naming seems various. On this issue, probably another, more generic discussion on alternative approaches may be needed to form several consensuses, as you said. I'm glad if someone do it ... however, it may be my task, of course.
As for "realtime control", it may be also another important aspect of additive synthesis (possibly related to several researches on "Real-time additive synthesis"). However, the number of controller simultaneously controlled by human is, so far fewer than the number of parameters on additive synthesis. Probably, several grouping of parameters or partials (as on Kawai K5 or Group additive synthesis[10][11][12]) may be needed, and ability of each implementations may various. Also it requires another sources and verification. Anyway, I added it to above categorization. thanks --Clusternote (talk) 18:08, 11 January 2012 (UTC)[reply]
Upper categorization
[new addition in 2011-01-11 18:00 (GMT)]
In addition to above internal categorizations, also upper categorization on widely accepted, more generic categorization system, may be also needed to clarify scope of article.
Keep in mind that trying to create strictly hierarchical taxonomies is inconsistent with the nature of human activity. For example: additive synthesis was one inspiration to Spectral Modelling, or you could say Spectral Modelling *uses* additive synthesis or *is a form of*, or *is an elaboration of*. However to flip the heirarchy and say that additive synthesis is a subset or (degenerate) form of SMS unnecessarily privileges SMS as some kind of super technique that "owns" additive synthesis (indeed some might like it that way, including your reference -- one of the inventors of SMS). The neutral way, in my view is to avoid imposing unnecessary taxonomies and hierarchies and prefer enumerating relationships, common usage, and chronologies. --Ross bencina (talk) 06:05, 14 January 2012 (UTC)[reply]
Thanks for your information. As for the above upper categorization, I didn't yet add exact semantics as you said "super", "subset", etc., because it may be varies for each tree. If SMS was POV, it may be ignoreable because it was not "widely accepted, more generic categorization". I also think that "imposing unnecessary taxonomies" should be avoid, as you said. --Clusternote (talk) 08:26, 15 January 2012 (UTC)[reply]
Historical trends of notions
[new addition at 2011-01-11 18:00 (GMT)]
Related to the definition and categorization, also historical transition (or trend) of notions (including precedents), may be needed to clarify the story. The section "implementations" is not yet enough organized, thus may be fit to this purpose. Though, I haven't intention to form a rigid structure which eliminate exceptions.
The following is my outline, as a starting point:
medieval (12th) to early 19th: prior to Fourier's theorem, early empirical notion (timbral additive ?) similar to "Steady-state, Harmonic additive synthesis", was seen on pipe organs with multiple ranks (multiple sets of pipes on which each set (rank) has different timbre, formed in 12th century).
19th to early 20th: Fourier's theorem became basis of earlier empirically known notion. Teleharmonium (c.1897) and Hammond organ (1934) using quasi-sinusoidal waves may be its implementations.
1910s–1930s: The early notion of "Time-varying additive synthesis" was implemented by several Russians (and probably others; see 120years.net) as manually drawn summation of sinusoidal waves on the films. Under the influence of them, later ANS synthesizer (1937-1957) was developed.
late 1950s–1960s: notion of "sound synthesis" and "computer music" was established. Earlier researches on "additive synthesis" for computer music were probably also established in this era. (ToDo: citations for earlier researches)
1970s–early 1980s: With the evolution of digital microelectronics including microprocessors, several additive synthesizers were implemented as digital hardware synthesizer. Bell Labs Digital Synthesizer (Alles Machine) was developed following earlier software experiments on Bell Labs (possibly non-Analysis base). Several later models including Fairlight CMI, Synclavier, had additive resynthesis feature on sampled sound (Analysis base).
mid 1980s–mid 1990s: most of consumer models without Analysis were released in this era.
mid 1990s–current: software synthesizer on PC appeared in late 1980s-early 1990s came to main stream.
instead of generic Fourier's theorem[2].
This implicit description also ignores preceding "analysis" phase on which Fourier transform (FFT or STFT) or "analysis filter-bank" are used to extract parameters from given audio signal. It should be improved. --Clusternote (talk) 14:07, 10 January 2012 (UTC)[reply]
Moreover, in my eyes, the section title "Theory" seems to be slightly too grandiose and possibly cause misleading. If the section mainly discussed on definition and not mentioned on any theoretical predictions on cause & result (explained by theory), it should be named "Definition". --Clusternote (talk) 18:59, 11 January 2012 (UTC)[reply]
Smith III, Julius O.; Serra, Xavier (2005), Proceedings of the International Computer Music Conference (ICMC-87, Tokyo), Computer Music Association, 1987., Center for Computer Research in Music and Acoustics (CCRMA), Stanford University {{citation}}: |chapter= ignored (help); External link in |chapterurl= (help); Missing or empty |title= (help); Unknown parameter |chapterurl= ignored (|chapter-url= suggested) (help) (online reprint)
Yee-King, Matthew (2007-01-15), "Music Technology Lecture 3, Sound Synthesis 2"(PDF), Lecture slide, Department of Informatics, University of Sussex{{citation}}: CS1 maint: date and year (link) Terminologies used on this article: "Harmonic Additive Synthesis (time invariant)", "Partial Addition", "Inharmonic Additive Synthesis (time invariant)", and "Time varying Additive Synthesis".
Interactive Swarm Orchestra (n.d.), Basic Synthesis Techniques, Interactive Swarm Orchestra / Interactive Swarm Space (a project realized as a collaboration between ICST of Zurich University of the Arts, and AILab of University of Zurich){{citation}}: CS1 maint: year (link) Terminologies used on this article: "Steady-State Additive Synthesis" and "Time-varying Additive Synthesis".
^ abB. Eaglestone and S. Oates (1990). "Proceedings of the 1990 International Computer Music Conference, Glasgow". Computer Music Association. {{cite journal}}: |chapter= ignored (help); Cite journal requires |journal= (help)
I would like to offer a brief review of the page as of January 12, 2012. I admit I have not digested the totality of this talk page. But the fate of the article as a whole is important and I think there needs to be some structural change and clarification of priorities.
I think a good way forward for this article would be to introduce an improved structure which asserts a heirarchy of information. Once an appropriate heirarchy is established I may have more to say on the fine detail. Certainly I agree that "time varying" is a widely understood term that is independent and unrelated to "real-time computing" (you can have a time varying signal generated by a non-real-time system, and vice versa)
Most importantly, the first section after the introduction should be "Theory". "Additive resynthesis" is a subtopic and best introduced later. Alternatively there could be a general overview section that includes reference to Analysis/resynthesis but I can't help but feel that Analysis/resynthesis should be a separate page -- since the idea of analysis/resynthesis goes well beyond Additive *synthesis* which should properly be defined as the combination of sinusoids, possibly with mention of some historical precursors.
With regards to the theory section, my opinion is that harmonic additive synthesis is more properly originated in Pythagoras' purported experiments with harmonic (integer) subdivision of strings. And could also be considered in in the context of inharmonic modes/resonances. Note that the presences of sinusoidal resonances ("modes") in a sounding body -- which is the basis of being able to synthesis these sounds using sums of sinusoids -- has, in general, nothing to do with harmonic series, and is more related to modal synthesis and modal decomposition of a sounding body. Certainly this kind of understanding of the harmonic series, modes, and of sound as a sum of sinusoids, pre-dates J. Fourier's work and is arguably more relevant.
Therefore I suggest that "Theory" should begin not the Fourier Series but by discussing the harmonic series [10], and perhaps of modes in non-harmonic sounding bodes. Given that this is a sound synthesis article, begining by relating it to acoustic theory rather than Fourier theory would make sense.
Further, with regard to Fourier theory: additive *synthesis* is not necessarily to be understood as being the same as Fourier Resynthesis (although the former may be implemented in terms of the latter, i.e. Rodet's FFT^-1). A resynthesis using the DFT is often a *phase vocoder* resynthesis.
I agree that the "Implementations" section is not neutral. I think precedence should be given to systems that mix sine waves. Combining arbitrary sounds in an additive manner (i.e. mixing) is in some sense analogous to additive synthesis but unless the process yields a single coherent fused timbre in is just mixing (plenty of synthesiser have multi-layer mixed timbres, and no one confuses these as additive). Pipe organs are not strictly additive synthesis -- the "Implementations" section appears to favor what I would call "Historical precursors" as if they truly are additive synthesis. I recommend that real additive synthesis devices are listed, and then perhaps have a section on "Historical precursors".
There should be a section that maps out the chronology of sinusoidal additive synthesis: from adding individual sine waves, to hybrid analysis-resynthesis approaches such as Spectral Modelling Synthesis (perhaps listed as a "related technique". Some thought needs to be given to whether and how to make reference to phase vocoder.
A nit pick: I would question whether the introduction should say "sinusoidal overtones" the usual general term is "partial", since overtone implies a fundamental frequency -- and in general, additive synthesis is just as applicable to synthesizing inharmonic timbres (e.g. gongs) as it is to synthesising harmonic tones. In any case, the introduction could just say "that creates timbre by explicitly adding sine waves (or sinusoids) together.
--Ross Bencina (talk)
I'm glad for your excellent review. Although several points are painful for me, your fourth paragraph mentioning Pythagoras and harmony is very interesting. For describing precedents of additive synthesis prior to the Fourier's theorem, several mentions on historical studies of harmonics, and clarification of relation between harmony and additive synthesis, seems to be necessary.
This article (Additive synthesis) in general, tends to biased to implementation side (especially by a electric engineer), and seems to lack more essential explanations in the broader viewpoints including acoustic engineering (sound waves as physical phenomenon) and psychoacoustics (harmonies and timbres as mental phenomenon). These broader viewpoints must provide both rationality of additive synthesis, and criticism on its potential and limits, in my opinion.
(please add your overall opinions under here, if possible)
Writing style
I'm not a subject-matter expert here, but I do know how to write for Wikipedia. This diff reverts typical changes to clean up language in articles that obscure matters for the "general reader". Simpler language that does the job concisely is actually the house style here. Charles Matthews (talk) 19:28, 12 January 2012 (UTC)[reply]
I agree that the diff introduces increasingly obscure language, however it also corrects falsehoods. Cleaning up the language to be black-and-white can only be appropriate if the facts are black and white. In this case I don't think they are. Ross bencina (talk) 19:54, 12 January 2012 (UTC)[reply]
We could cut out all the stuff in the middle (but I wouldn't recommend it). It exists to generalize from the Fourier series representation used for harmonic tones and to take it to the simplest, most general expression of additive synthesis (adding sinusoids with arbitrary frequency and amplitude). 70.109.183.99 (talk) 20:55, 12 January 2012 (UTC)[reply]
Could you explain the intended meaning of "quaai-periodic" there? In mathematics it is ambiguous, but a straight traditional Fourier series does not represent anything quasi-periodic, for sure. Charles Matthews (talk) 21:29, 12 January 2012 (UTC)[reply]
There is a problem with nomenclature overlapping with other disciplines. It is most closely related to Almost periodic function. It is a function where some periodicity is evident; one period or cycle looks nearly indistinguishable from the adjacent periods but may look very different from periods that are far away. Other terms used have been "quasi-harmonic", but that alludes more to the frequency-domain nature of the tone rather than the time-domain. Many musical instruments have notes that, at least after the initial onset, are quasi-periodic or quasi-harmonic. As far as the Fourier series, it means that the ak and bk coefficients are not constant (if they were constant, the result would be perfectly periodic, of course) but are functions of time that are relatively slowly varying, like a synthesizer ADSR envelope but more general than ADSR. 70.109.183.99 (talk) 21:52, 12 January 2012 (UTC)[reply]
Fixed-waveform vs time-varying
There has been some confusion between "real-time" (i.e. something capable of generating sound in real time, such as a computer program), and "time-varying", which in the context of additive synthesis means that the partials change amplitude and/or frequency over time. Clearly given these definitions it is possible to have both time-varying non-real-time synthesis and also non-time-varying real-time synthesis. I hope this entry provides sufficient explanation as to why I have replaced real-time with time-varying in almost all cases in this article. It should be noted that whether or not something is real-time, while of technical interest, is of little substantial relevance to the consideration of a synthesis techniques on its merits to produce particular timbres.
The fixed waveform / time-varying classification is supported by Roads 1996:
Additive synthesis is a class of sound synthesis techniques based on the summation of elementary waveforms to create a more complex waveform. Additive synthesis is one of the oldest and most heavily researched synthesis techniques. This section starts with a brief history of additive synthesis and explains its fixed-waveform and time-varying manifestations. (Curtis Roads, The Computer Music Tutorial, p134. MIT Press 1996)
You're completely correct, Ross, but there is some history of the first real-time additive synthesis machine (the Bell-Labs machine?) and there is a concept of real-time synthesis as opposed to creating a sound file in non-real-time and playing that back later (at the proper speed). This is part of the early history of additive synthesis. At first most of us couldn't do it with a real-time device where you would hit a key and out would come the sound. So it is appropriate to have some reference to the concept and I wouldn't even object to calling a Hammond B3 a sorta crude "real-time" additive device and I wouldn't object to calling an organ a precursor to additive synthesis.
Essentially, the only difference between "static" and "time-varying" is that the rk and φk are constant (for harmonic) and that rk and fk are constant for for non-harmonic.
I believe that one thing that may have been hanging Clusternote up is that if any of these parameters vary because someone turns a knob (in response to a MIDI Control message) and if the synthesis responds immediately, that does have to be real-time. But the concepts are different. 70.109.183.99 (talk) 20:46, 12 January 2012 (UTC)[reply]
Criticism of the introductory paragraph
At the time of writing the introductory paragraph is:
Additive synthesis is a sound synthesis technique, based on the mathematics of Fourier series. It creates timbre by adding overtones together.
This has a few problems:
- The first phrase "Additive synthesis is a sound synthesis technique" implies that we are discussing a single technique. This is arguable, since it is possibly a family of techniques. Especially if you are going to include digital, analog, pipe organ, harmonic, inharmonic, fixed-waveform and time varying. Note that the Roads quote I included elsewhere on this page defines it as a family of techniques. None the less, if it is assumed to be an overarching category for all these techniques, then the second phrase of the first sentence is problematic:
- The second phrase "based on the mathematics of Fourier series" refers strictly to *harmonic* additive synthesis (as per the Theory section). Therefore it would be better to say "inspired by the mathematics of Fourier series" (influenced perhaps?) or some similar construction. I think it used to have something softer like this.
- In a similar vein "by adding overtones together" is also a reference to harmonic additive synthesis since overtones (at least as defined on the linked page) are usually understood to be harmonics, not just sinusoidal partials. The second paragraph is more accurate and complete and addresses these issues.
Overall I think this would be better:
Additive synthesis is a family of sound synthesis techniques where timbre is created by adding multiple components together. In one common form, components are sine waves arranged in a harmonic series.
So change it, Ross. If I did, I might get slapped down. I do not think that (except in the context of group additive synthesis) that there should be anything other than sinusoids being added in the lede description. How about?:
The problem is, Charles, that additive synthesis is about adding sinusoids together. As best as I can tell, it was not meant to be about adding square waves or triangle waveforms or Walsh functions or some other set of basis functions together. If the lede does not make that specific point (we're adding sine waves together), I believe it will be inaccurate. But the sinusoids do not have to be harmonically related, that is also important in the fundamental description of what additive synthesis is. 70.109.183.99 (talk) 22:02, 12 January 2012 (UTC)[reply]
If you can reference a definition of "additive synthesis" which really does say that (i.e. linear combinations of sinusoidal waves that may have incommensurable frequencies), then fine. Wikipedia does ultimately work on definitions that are "textbook", and if so those will be verifiable. The lede needs to cut through to essentials, so definitions are key, certainly. If the definition is right, then saying "is a technique", "is a family of techniques", "is one of a number of techniques", "is any technique" that does X, is the sort of nuance which is not really fundamental to explaining the main deal. It can go lower down the article. Charles Matthews (talk) 22:33, 12 January 2012 (UTC)[reply]
Additive synthesis is evidently the first technique widely used for analysis and synthesis of audio in computer music. It was inspired directly by Fourier's theorem (which followed Daniel Bernoulli's insights) which states that any sound s(t) can be expressed mathematically as a sum of sinusoids. The term "additive synthesis" refers to sound being formed by adding together many sinusoidal components modulated by relatively slowly varying amplitude and frequency envelopes.
So, Charles, what might you want to do with that? I think "inspired by" is better than "based on", but I also think that "Fourier series" is better than "Fourier's theorem". There is an inverse FFT method of doing additive synthesis (and it's messy), but what it is, essentially, is adding up a finite number of discrete sinusoidal components of specified amplitude and frequency (and inverse Fourier transform is normally more than that). The parameters going in is a finite set of frequency/amplitude pairs and sound comes out. 71.169.180.195 (talk) 05:35, 13 January 2012 (UTC)[reply]
What I'd want to do with that is to add a third para in the lede explaining what is going on in terms of modulation. This would be a good style for us: so-called "concentric": first para very crip, second para (which I have just tweaked) more technical. Then the third para amplifying what is said above it in terms that an expert could accept, and alluding to typical examples. That would be good, I think. Charles Matthews (talk) 08:22, 13 January 2012 (UTC)[reply]
I think it would be useful to acknowledge that there appear to be two separate concepts that are commonly referred to as "Additive Synthesis." (1) the one concerned only with sinusoids, supported by 70.109.183.99 and Julius Smith. It is also a view I have supported elsewhere on this talk page. (2) there is another usage related to a general synthesis "paradigm" of mixing simple components to create complex timbres -- this does not strictly require sinusoids, just spectral fusion or perceptual coherence of the result. In this sense it is additive (mixing) in contrast to subtractive (filtering). This latter definition is hinted at by the following textbook definitions:
Computer sound synthesis for music general falls into one or more of four basic categories: (1) *Additive synthesis models*, in which elemental sound components (such as sine waves) are added together in time-varying ways. Elements of Computer Music, F. Richard Moore, Prentice Hall, 1990. p 16.
The basic notion of additive sound synthesis is that complex musical sounds can be created by adding together multiple sound components, each of which is relatively simple. Such components are typically not perceived individually, but each component contributes materially to the overall quality of the resulting sound.
The greatest advantage of the additive method of synthesis is that it rests on the very well developed theoretical framework of Fourier analysis. For this reason, additive synthesis is sometimes called *Fourier synthesis*, though that description is more restrictive than the more general notion of additive synthesis.Ibid. pp. 207-208
*Additive synthesis* is a class of sound synthesis techniques based on the summation of elementary waveforms to create a more complex waveform. The Computer Music Tutorial, Curtis Roads, MIT Press 1995. p 134. [While most of Roads' chapter refers to sums of sinusoids it begins with the pipe organ and ends with Walsh functions.]
In my view this page could safely take the "sum of sinusoids" definition as the principal one. This agrees with previous comments by 71.169.180.195. For completeness, a "Broader interpretations of Additive Synthesis" section could be added, quoting Roads above and listing related methods that don't involve sinusoids (such as Walsh functions, pipe organs, tone wheels etc). Does this sound OK Charles?
Ross bencina (talk) 10:30, 13 January 2012 (UTC)[reply]
In light of the above I have edited the lead to focus on sine waves. I have added a "Broader definitions of additive synthesis" section. This text could perhaps appear instead in the lead. Thoughts?
Ross bencina (talk) 16:18, 13 January 2012 (UTC)[reply]
I'm okay with all of this, Ross. I have trouble, historically and conceptually, equating this Walsh function synthesis or Wavetable synthesis to Additive synthesis. Heck, if you stretch this broadened definition enough, you can say that Sampling synthesis is a form of Additive synthesis, since it adds up a bunch of Kronecker deltas:
Fully agree. Both the Roads and Moore quotes above reflect attempts to use Additive Synthesis as a domain-level taxonomical division. In (Roads) it is used as a chapter title to conveniently group somewhat related techniques. In (Moore) it is listed as one of four basic categories. In both cases "Component summation methods" would have been better. None the less I think it is an established alternative usage and it should be addressed by the article to avoid confusion.
Question: should we move the content of "Broader definitions of additive synthesis" to the lead?
I dunno. If it were up to me, I would start out with the principle definition of Additive synthesis being the explicit addition of a finite set of sinusoids of arbitrary and specified amplitude and frequency. This would be the case whether it's real-time or not. If it's not real-time, it creates a "sample" that can be played back later. Then I might differentiate between harmonic and not-necessarily-harmonic additive synthesis. In the harmonic case, I would connect the concept to Fourier series. Then I would say there are some other less-direct ways of doing this, such as the inverse FFT, and wavetable synthesis, the latter of which would be only for the harmonic case (unless you detune the harmonics by rapidly changing the phase, which is an arcane detail that shouldn't go into the article).
Then maybe I would extend the concept out to other basis functions like Walsh, but emphasize that virtually all of the time when "additive synthesis" is referred to, it is meant to be about adding sinusoids. Then maybe go into all the sinusoidal modeling and analysis/resynthesis STFT stuff. That would cover nearly all of the academic stuff and what would be left would be the historical stuff about what they built long ago and what commercial products that do it or have done it. I really don't have all the ideas down for how to lay out the article.
On that basis I think we should leave things as they are at the moment (i.e. a separate section late in the document about more inclusive definitions.) Calling a Walsh function additive synthesis is categorically different from the other "advanced" methods such as SMS -- as such it is really just a style question of whether to include "Definition 2" in the lede. Perhaps with a single sentence "Additive synthesis has also been used as a general term to refer to the class of synthesis techniques that sum multiple elementary components." The whole lede section still needs a re-write. I may attempt it later. --Ross bencina (talk) 06:24, 14 January 2012 (UTC)[reply]
Inline Template:TeX vs HTML and what truly tidies up.
Hi Chris Johnson,
I've been editing Wikipedia since maybe 2004 or 2005 (you'll have to take my word for it, because I won't show you the editing history) and I really wish there was just one way to do math on the pages, but there has always been this discussion/debate about what looks better. In addition, there is a new math mark up construct {{math|''f''(''x'') {{=}} ''x''<sup>2</sup>}} or f(x) = x2. I am not sure what this latest construct gets us, but it's there.
I would normally want to use precisely the same construct for inline math as I would for an equation that stands alone on a line. That would normally mean LaTeX, like , but for some reason that is discouraged here. In addition, if there is no \ {space} character on a line that can be reduced in size, the inline LaTeX is reduced to display differently than one might expect. Sometimes even, the appearance is identical to what it would look like if it were inline HTML. Here it is without the backslash: , just like f(x) = x2.
After much back-and-forth with other technical and math editors (User:Michael Hardy comes to mind), I finally came to the conclusion that the best and most accepted formatting decision is to use <math> ... </math> for separate equations that live on their own lines, and I put in a backslash-space in there to make sure it renders like LaTeX, and inline with the text body, I do everything I can to markup with HTML constructs. That looks the most consistent and readable and is essentially what all these other math editors told me to do. Do you think the format changes you made really makes the math look better? I really do not.
Also, if we do not connect this additive synthesis theory to the concept of Fourier series, we can cut out all that crap and just say: This is what Additive synthesis with sinusoids is:
or
where
That is all that it is. And maybe the Theory part should be stripped to that. But if you wanna relate this to Fourier series, I do not see a way that is as complete and concise as what we had to start with. So, what do you guys think? Including Clusternote. You get to contribute to the article too, as long as you do not try to take it over again and replace correct mathematics that you don't understand with incorrect mathematics that you think you understand. And
is incorrect mathematics since it doesn't work unless fk[n] is constant w.r.t. n.
Also, for historical reasons, we should revisit the real-time vs. non-real-time thing again. That Bell Labs synth was the first real-time (in the full meaning of the term) additive synthesizer, if I recall correctly. Before that, we would write computer programs that would run at maybe 5 or 10 times slower than real time (so it would take 5 or 10 minutes to create 1 minute of music) and would compute a soundfile that would be written to hard disk and reproduced at a later time. That is the essential difference between real-time and not. A real-time synth is one where you hit the key and the note is synthesized on the spot. That also means if you slide a fader that is supposed to affect the note, you will hear the effect immediately. This is about being live and is related to time-varying but is not the same thing. But there is an historical issue about what is real-time additive synthesis and what is not. And that should be in the article. 71.169.180.195 (talk) 18:59, 13 January 2012 (UTC)[reply]
I'm happy to change the maths formatting to whatever is the accepted standard. The inline maths before I edited it was just italic sans-serif text produced with with the standard double quotes (''x+y'' = x+y). I changed that to using the math tag, which for the simple inline equations on the page displays (on my computer) as an italic serif HTML font rather than as LaTeX (<math>x+y</math> = ). I'm quite happy to change this to the new HTML math markup style ({{math|''x'' + ''y''}} = x + y) if that's the accepted way of doing things -- or were you suggesting going back to the sans-serif math without the new math markup tag (x+y)? (To my eyes, the italic serif font produced by inline <math> or the new HTML markup looks nicer than the italic sans-serif font that was there before my edits, but it's going to be browser/OS dependent so I can well believe it looks worse elsewhere.)
On the subject of what is required in the theory section, we could indeed strip it down to the three equations that you have above, perhaps then adding that if everything is time-invariant, this is equivalent to the form, which is a Fourier series. However, if additive synthesis is presented in the introduction in the context of Fourier/harmonics/etc., then we should probably start with the equations in a Fourier-like form and show that they can be extended to inharmonic/time-varying frequency, as the page is currently (though I'm happy to reduce the amount of detail considerably compared to what is there presently). Chrisjohnson (talk) 15:45, 14 January 2012 (UTC)[reply]
I'd be happy to use {{math}} in inline equations, I think it flows better with the text than <math>. Makes math look less scary, so to say. But everyone appears to use <math> currently so... Perhaps some gnome will do the conversion some day. Thanks already. Olli Niemitalo (talk) 01:04, 17 January 2012 (UTC)[reply]
How about having the following kinds of equations for continuous-time output signal y(t), each in the most simple form, in a Theoretical background section (from the most easy to understand and simple to the most complex)?
1. Continuous-time, harmonic partial frequencies, constant partial phase offsets (phase at time zero), constant partial amplitudes.
Then the body of the current theory section could be moved to a separate Discrete-time equations section. I've began to think that maybe we should not cripple (by removing equations) the rather solid block of discrete equations too much. It's a nice reference. Olli Niemitalo (talk) 18:07, 14 January 2012 (UTC)[reply]
Okay, can I ask that someone who is more than an IP (IPs cannot upload images to the en Wikipedia) create a simple drawing similar to Figure 22.1 at the bottom of this reference? Please leave out the noise source and, for consistency, use r1(t) instead of A1(t), etc. And please label the output of the big summer y(t).
We could use that simple diagram to start with. Then we can start with the simplest, continuous-time, and general (harmonic or inharmonic) additive synthesis equations. Then we can work this the other way, first turn them into discrete-time counterparts, then, for the harmonic synthesis case make these more specific toward harmonic synthesis, and then relate it to Fourier series. Even though the electrical engineers among us might like to see these equations in the more compact form using the complex exponential, I really think that we should keep this real, so we would have cosine and sine terms when we compare to Fourier series. One advantage of doing it this way is that we can get rid of the absorbing the change of phase into the instantaneous frequency term step that has been hanging up some unidentified persons.
So would someone want to create an PNG or SVG graphic like the one in the JOS reference? Clusternote, would you like to do that and contribute usefully to the article? That is something I cannot do. 71.169.180.195 (talk) 21:52, 14 January 2012 (UTC)[reply]
When drawings are needed, it should be drawn by the someone who think it is absolutely needed, to express his/her own original intention. Most users are usually doing so. If someone can't upload his/her drawing on Wikipedia for several reason, still he/her can transfer its drawing via other upload service with the help of other users, as we helped someone a few days ago. --Clusternote (talk) 06:55, 16 January 2012 (UTC)[reply]
Review of assertions in Wavetable Synthesis section
I just deleted the following unattributed comment from Wavetable Synthesis section:
wavetable synthesis requires just as much computation as additive but transfers much of the computation to a pre-synthesis step.
I would like to point out that this claim is incorrect. Wavetable synthesis only requires pre-computation of a relatively small set of single-cycle waveforms ("breakpoint waveforms" if you like). These can then be looped and slowly crossfaded to ramp between sets of partial amplitudes. A cross fade from one wavetable to another is the same as all of the partials fading from one set of amplitudes to another. The Beauchamp reference elsewhere provides concrete examples I think.
On the other hand, it could be claimed that wavetable synthesis is *possibly* less space-efficient than oscillator bank additive synthesis. However, the article makes no claims about space efficiency.
--Ross bencina (talk) 10:53, 14 January 2012 (UTC)[reply]
Oh, you're probably right on all counts, Ross. Given the same frequency-domain additive synthesis parameters (the rk(t) and φk(t) envelope functions, where φk(t) is sorta related to the frequency, fk(t), of the harmonic partial - it is possible in wavetable synthesis to detune it a little from the perfectly harmonic; k f0), when t is fixed to some value, both (harmonic) additive synthesis and wavetable synthesis do a form of inverse Fourier transform to create a time-domain waveform. It's just that wavetable does that in advance of the MIDI Note-On event and straight additive does it immediately after the MIDI Note-On event. They both have to do that computation, but for wavetable, the time-domain data is loaded up and ready to rock-n-roll when the key is pressed.
If the breakpoints are all at the same times for all rk(t) and φk(t) envelopes, and if your wavetable isn't "oversampled" (that is the number of wavetable points is about equal to 2K where 1 ≤ k ≤ K), then the amount of space required by the two methods is about equal. But if the breakpoints in the rk(t) and φk(t) envelopes do not fall on the same times (for different k), then wavetable synthesis is less space-efficient because, for wavetable, once one rk(t) requires a breakpoint, then all of the rk(t) functions get a breakpoint at that same time. So, because if this added restriction, the breakpoints aren't as optimally determined. But those are the only two reasons I can think of for why the wavetable data would take significantly more space than the straight additive data; one is if the wavetable is oversampled and the number of wavetable points greatly exceeds 2K and the other is the less than optimal breakpoint locations because the breakpoints for all harmonic envelopes must occur coincidently.71.169.180.195 (talk) 19:01, 14 January 2012 (UTC)[reply]
Audio samples
I note that this page has a {{Reqaudio}} requested recordings tag. I'm happy to synthesise some (and provide the C++ source for them); I was thinking of making one sound with harmonic partials of time-varying amplitude (showing how the timbre varies for a fixed fundamental frequency), and one more abstract sound with many evolving inharmonic time-dependent-frequency partials. Any thoughts? Chrisjohnson (talk) 16:01, 14 January 2012 (UTC)[reply]
Would be neat to have also graphical presentations of those sound samples. Spectrograms maybe, or envelopes of the partials drawn in different colors in an amplitude vs. time plot. Something illustrative, not too pedantic. Olli Niemitalo (talk) 16:47, 14 January 2012 (UTC)[reply]
That looks great! It's fine as it is if we drop these in different places along the story, but if we want a gallery (with graphs on top and audio below each one) then the figure would need to be compacted horizontally so that a few will fit on the same row. Dunno which it is yet. Olli Niemitalo (talk) 10:59, 15 January 2012 (UTC)[reply]
Inharmonic example added above, and the images now reduced in width. I can't figure out how to get the images on top of the audio samples while retaining the 'play' button (though there is a way of doing it if the audio is a linked just from a piece of text, rather than from a button). Chrisjohnson (talk) 01:22, 16 January 2012 (UTC)[reply]
I understand that these examples are intended to be illustrative of concepts in the theory section -- for this they are excellent. But would it also be possible to have at least one example that sounds like a musical instrument? (a piano tone, a vocal sound, or perhaps a gong or bell?) It doesn't have to be a resynthesis (probably it shouldn't be) but it should at least give the reader the idea that additive synthesis is useful for synthesising *common* *complex* musical timbres. Such an example could be used in the introduction before the theory section, and titled something like "Additive synthesis of a Piano-like tone". Just an idea. Ross bencina (talk) 05:03, 16 January 2012 (UTC)[reply]
Good idea - I'll put together such an example. I looked for a less abstract example in (popular) recorded music that could be used - but couldn't find anything that was both free and definitely produced by additive synthesis. If anyone can think of such an example, it would be nice to use that: it would illustrate the significance/history of additive synthesis in a way which samples that are created just for the wikipedia page cannot. Chrisjohnson (talk) 11:31, 16 January 2012 (UTC)[reply]
I think it is best if both visually interesting synthesis sample by Chris, and other more natural samples, were shown on article, if possibly.
Also I remember one of earliest gong sound emulation was done on Trautonium in 1942 at the latest. Possibly it may be one of the earliest inharmonic additive synthesis. --Clusternote (talk) 05:32, 16 January 2012 (UTC) P.S. Also it should be added to timeline section. (possibly later "subharmonic synthesis" used on Mixture Trautonium (1952) and Subharchord (1960s), may be more appropriate) --Clusternote (talk) 06:07, 16 January 2012 (UTC)[reply]
Theory section needs a drawing like in the JOS article.
Clusternote, this pair of edits were useful and good, which is why they remain.
Please consider straightening out your own understanding about what is happening regarding the argument of a sin() or cos() function and what instantaneous frequency is, before attempting to transmit that understanding to others via a publicly-consumed article in the WP:main namespace. The talk page is good for that, but you'll have to try to form well-defined questions for the mathematicians and engineers to answer and then you need to listen. You can make contributions to the article that has value, but they won't be of value if they promulgate misunderstanding rather than understanding.
Now, what this article really needs, in my opinion, is a simple drawing that shows a few simple oscillators, each with independent control of frequency and amplitude, and the outputs of each of those oscillators summed into a big summer that has output y(t). I cannot upload this graphic. is there someone who would be willing to do it? I asked Clusternote if he/she might want to contribute to the article in that positive manner, but I suspect that this request was declined. 71.169.180.195 (talk) 19:07, 15 January 2012 (UTC)[reply]
As already discussed, we needs appropriate supplemental explanations on the discrete equations shown on the Theory section (its explanation is still inconsistent and lacks rationality), for "general readers" who studies mathematics in continuous form.
Supplemental note for section "Inharmonic partials" using continuous form
As we already discussed on above, "Additive synthesis#Inharmonic partials" section which described on too specific details of discrete implementation needs supplemental explanation using continues form for readers familiar with continuous form. Also it is needed for additional verification on complicated description using discrete form on the section. I've added below supplemental note on Additive synthesis#Footnotes section.
For readers not familiar with above notations, descriptions using these are possibly not easy to understand at a glance. The following is a try to provide supplemental explanations for these readers.
In continuous form, a wave (correspond to a partial on the 2nd equation) is expressed as
However, an ill-mannered IP user seems to start his faithless revert-war on footnote without preceding discussion, and even start personal attack on talk page and edit summary field using his poor English. Probably, he still not understand above description is just an outline using continuous form corresponding to discrete form on the section.
With our intelligence, how to handle this problematic IP user ? For this specific issue, I expect advices from third person who isn't concerned on this issue (i.e. other than me and IP user).
best regards, --Clusternote (talk) 02:08, 16 January 2012 (UTC)[reply]
In my opinion, IP user provided sufficient reason for his deletion in the edit summary. Clusternote: the guidelines ask you to refrain from personal attacks (criticising "poor English" in this case). So far everyone has refrained from criticising your poor English and I think you should give other people this same respect. Ross bencina (talk) 05:11, 16 January 2012 (UTC)[reply]
Why you think he provided sufficient reason ? We needs supplemental explanations on corresponding section for "general readers" who studies physics and mathematics in continuous form. If you found any defeats on supplemental notes, it should be improved instead of reverted. I expect your honest criticism on above notes.
By the way: My ironical expression "poorly English" means "inappropriate words" shown on his edit summary. (also he repeatedly cause personal attacking in the past). As you recognize, I also recognize rather my English is poor : ). sincerely, --Clusternote (talk) 05:55, 16 January 2012 (UTC)[reply]
Or, if you do not want to hear any criticism from me, would you allow me to spell out to you exactly what this instantaneous phase and instantaneous frequency is about? 71.169.180.195 (talk) 06:26, 16 January 2012 (UTC)[reply]
If it becomes apparent to other editors (like Chris or Olli or Ross) that the effort and time required to disassemble your footnote exceeds what they would be able to provide, and since the time they have to work on the article likely is already limited, if it becomes just too difficult and inefficacious to get you to understand, can you accept that the footnote is not ready for the main namespace? I wasn't the only editor to remove your contribs to the article and this is the first time I did remove it since the previous "edit war", and since there has been robust participation in the article by some very knowledgeable and productive editors. 71.169.180.195 (talk) 06:26, 16 January 2012 (UTC)[reply]
IP user 71.169.180.195, you didn't explained any about continuous form other than personal attacking on your past posts. If you can't explained it in continuous form, you shouldn't discuss on it. --Clusternote (talk) 07:08, 16 January 2012 (UTC)[reply]
Clusternote, do you think that the other editors agree with your evaluation of "problematic". I think you deem yourself as "not problematic". Do you think the other editors agree with your evaluation of "not problematic"? 71.169.180.195 (talk) 06:26, 16 January 2012 (UTC)[reply]
Then, how should we simply explain that section's too detailed implementation specific description to generic users ? Essentially, that section merely describes details of below equation (below is tail of above supplemental note).
Second equation shown on the top of section can be rewritten using above result, as following:
The last form matches with the "general expression" shown on the tail of section.
However, even after recent improvements, that section still seems complicated and hard to grasp its intention, without any supplemental note. --Clusternote (talk) 10:02, 16 January 2012 (UTC)[reply]
About instantaneous phase and instantaneous frequency, in the context of additive synthesis equations, we cannot equate our definitions of time-dependent phase and time-dependent frequency to those definitions because those are non-local and we have local definitions. Instantaneous phase of a real-valued signal is obtained by introducing an imaginary component that is the Hilbert transform of the real signal, and by taking the argument of the resulting complex number. Hilbert transform is obtained by convolution with the Hilbert kernel, which has no compact support, hence non-locality. In other words, instantaneous phase of a non-static sinusoid, at current time, will depend on the unknown future, indefinitely, and that's not what's happening in our synthesis equations. So it might be better to stick to the different names or to describe the difference, or both. Olli Niemitalo (talk) 11:00, 16 January 2012 (UTC) Edit: This applies to the discrete equations. To the continuous equations, I'm not sure! Olli Niemitalo (talk) 14:16, 16 January 2012 (UTC)[reply]
Actually, Olli, I'm not sure I agree with you. First, I do not think that there is much difference in outcome between the continuous-time case and the discrete-time case. Second, as long as the envelopes rk(t) are bandlimited sufficiently, the Hilbert transform of
At this point there is agreement with how we understand instantaneous frequency from the POV of the analytic signal. The instantaneous frequency of a continuous-time sinusoid is simply and always the derivative of the argument of the sin or cosine function w.r.t. time and it need no definition or relationship with the analytic signal when only real sinusoids are involved. 71.169.180.195 (talk) 02:01, 17 January 2012 (UTC)[reply]
I know you wish for that to be true, Cluster. But it isn't. For example, the expression has no meaning. (The reason is that you're mixing discrete-time notions with continuous-time notions ad hoc in the same equation. The only way to relate continuous-time notions to discrete-time, is via the sampling theorem.) Your supplemental note is dead-in-the-water, right from the beginning. 71.169.180.195 (talk) 04:02, 17 January 2012 (UTC)[reply]
Is it sure ? The following two expressions are probably correct.
Those two equations are correct. And they are equations that live solely in the continuous-time domain. So, in your note, can you tell me what domain φk[n] is in? And what meaning there is to (d/dt)φk[n]? (d/dt)φk(t) does have meaning, but (d/dt)φk[n] does not and if you want to convert φk(t) to φk[n] or the back again, you need to consider the sampling theorem. BTW, the \textstyle TeX command does not appear to do anything. I don't know why you like putting it in. Use HTML or LaTeX, whichever you like, but try to keep the equations clean and consistent in style and use. Makes it easier for others to see what you're saying and, if they respond, to copy and edit your equations in response. Glad you're playing nice now. 71.169.180.195 (talk) 04:35, 17 January 2012 (UTC)[reply]
71.169.180.195, well, I think I'm still going to feel uneasy about it unless we mention the band-limiting requirement for and also for (thus allowing the two definitions of instantaneous phase and instantaneous frequency to almost meet, and also pretty much hides the fact that is "centered" in-between times of sample and ). — Preceding unsigned comment added by Olli Niemitalo (talk • contribs) 03:14, 17 January 2012 (UTC)[reply]
Well, Olli, they don't do that so much in our communication theory and signal processing textbooks. It's because no envelope that is time limited can also be bandlimited. All we need to worry about is that these envelopes are sufficiently bandlimited so that they don't spill into the negative frequencies. That Hilbert transform and analytic signal relationship is actually an approximation, but a very good one. It essentially means that we can think of those envelopes as constant. If rk(t) does not vary too fast, it's variance does not affect the instantaneous frequency. Now when you apply the same reasoning to ak(t) and bk(t), then we start to get into a little problem because if they do not vary in tandem, then the effective φk(t) varies and we *know* variation of phase causes detuning to the instantaneous frequency. This is essentially why, if starting from the Fourier series perspective (with the ak(t) and bk(t) envelopes) that you have to have some way of absorbing the change of φk(t) into the instantaneous frequency. But if we don't start with that and don't even fiddle with
and start directly out of
we might be able to totally ditch the "absorb the change of phase into the instantaneous frequency" thing. And, I imagine if we do, Clusternote will want to take credit for it (and maybe some credit for motivating the issue is due him). Note that it is φknotφk[n]. I am contemplating how to do that succinctly and considering pedagogy. This is why I started that Fourier series section. I think, if we do this right, that section is the only place we should see ak(t) and bk(t), both sines and cosines at the same frequencies. After disposing of that, then the remainder should always be about a modulating amplitude rk(t) and either a modulating phase φk(t) with harmonic frequencies k f0 or an inharmonic and possibly time-variant frequency fk(t) with a constant phase offset φk. 71.169.180.195 (talk) 04:02, 17 January 2012 (UTC)[reply]
Probably I roughly grasped intention of your clear proposal. Moreover, direction of recent improvements on that section seem almost correct, in my eyes. The thing I can't grasp yet it is, why we should use still this specific implementation which seems slightly hard to rationally explained, as example on Wikipedia. Possibly, is it based on a defact standard code on that field ? (If so, probably we may be able to find several sources other than Smith III ...) --Clusternote (talk) 12:46, 16 January 2012 (UTC)[reply]
P.S. I understand that you are talking about signal processing using complex form. It is not my specialty, however, your above kind comment is very helpful for me to improve my understandings on these equations. I'm sorry for trouble you. sincerely, --Clusternote (talk) 17:44, 16 January 2012 (UTC)[reply]
P.S.2 I'm happy if I can clearly identify the definitions of time-dependent phase and time-dependent frequency and also local definitions used on equations described on that section. If possible ... --Clusternote (talk) 18:24, 16 January 2012 (UTC)[reply]
Wikipedia is used as a source of information by all kinds of people. While the general reader (say, a musician) might not be interested, the discrete equations may be useful for those who wish to create their own additive synthesizers (say, a computer programmer), or interesting to someone who wishes to know more about how their inner workings can be described mathematically (say, a "nerdy" person). Clusternote, perhaps it would be more helpful, as compared to you writing a footnote, if you would point out here on the talk page what exactly it is that you find disagreeable about the discrete equations. We can then respond by 1) clarifying the article and 2) correcting any true defects. Olli Niemitalo (talk) 14:16, 16 January 2012 (UTC)[reply]
Yet not clarified points
Why this specific implementation is selected as a sample on article ?
It seems share several with Smith III's articles (first one was originally added by me), however, it is still nearly single source state. To meets notability, we should add yet another reliable sources.
What theory is underlying on this specific implementations ?
Until early this month, I assumed this specific implementation (and although more generic definition I assumed) were probably related to Analysis phase, and based on theory of signal analysis and digital signal processing. However at now, we seems almost left Analysis as out of scope for this article.
Where can we find theory on the section ? and where is its reliable sources ?
Normally, merely a formula modification which lacks clear explanations on purposes, merits, assumptions and its proofs, is not called theory. It is simply called "a formula modification".
Also, we needs short general explanations outlining what is described on discrete section, for generic users (i.e. other than implementers) to help their understandings on these discrete equations. It may be probably described in a few equations in continuous form.
Note: on this year's earliest revision, this specific implementation was called "Theory", and it was described using instantaneous phase and instantaneous frequency. As a result, it had been incorrectly described for 4 years, until a few weeks ago. I'm glad for drastically improvements during last a few weeks !