Skip to content

Commit d86b273

Browse files
committed
Fix format
Signed-off-by: Taym <haddadi.taym@gmail.com>
1 parent 00f249e commit d86b273

File tree

4 files changed

+42
-42
lines changed

4 files changed

+42
-42
lines changed

html5ever/src/tokenizer/char_ref/mod.rs

+11-11
Original file line numberDiff line numberDiff line change
@@ -115,7 +115,7 @@ impl CharRefTokenizer {
115115
pub(super) fn step<Sink: TokenSink>(
116116
&mut self,
117117
tokenizer: &mut Tokenizer<Sink>,
118-
input: & BufferQueue,
118+
input: &BufferQueue,
119119
) -> Status {
120120
if self.result.is_some() {
121121
return Done;
@@ -135,7 +135,7 @@ impl CharRefTokenizer {
135135
fn do_begin<Sink: TokenSink>(
136136
&mut self,
137137
tokenizer: &mut Tokenizer<Sink>,
138-
input: & BufferQueue,
138+
input: &BufferQueue,
139139
) -> Status {
140140
match unwrap_or_return!(tokenizer.peek(input), Stuck) {
141141
'a'..='z' | 'A'..='Z' | '0'..='9' => {
@@ -156,7 +156,7 @@ impl CharRefTokenizer {
156156
fn do_octothorpe<Sink: TokenSink>(
157157
&mut self,
158158
tokenizer: &mut Tokenizer<Sink>,
159-
input: & BufferQueue,
159+
input: &BufferQueue,
160160
) -> Status {
161161
let c = unwrap_or_return!(tokenizer.peek(input), Stuck);
162162
match c {
@@ -177,7 +177,7 @@ impl CharRefTokenizer {
177177
fn do_numeric<Sink: TokenSink>(
178178
&mut self,
179179
tokenizer: &mut Tokenizer<Sink>,
180-
input: & BufferQueue,
180+
input: &BufferQueue,
181181
base: u32,
182182
) -> Status {
183183
let c = unwrap_or_return!(tokenizer.peek(input), Stuck);
@@ -207,7 +207,7 @@ impl CharRefTokenizer {
207207
fn do_numeric_semicolon<Sink: TokenSink>(
208208
&mut self,
209209
tokenizer: &mut Tokenizer<Sink>,
210-
input: & BufferQueue,
210+
input: &BufferQueue,
211211
) -> Status {
212212
match unwrap_or_return!(tokenizer.peek(input), Stuck) {
213213
';' => tokenizer.discard_char(input),
@@ -221,7 +221,7 @@ impl CharRefTokenizer {
221221
fn unconsume_numeric<Sink: TokenSink>(
222222
&mut self,
223223
tokenizer: &mut Tokenizer<Sink>,
224-
input: & BufferQueue,
224+
input: &BufferQueue,
225225
) -> Status {
226226
let mut unconsume = StrTendril::from_char('#');
227227
if let Some(c) = self.hex_marker {
@@ -270,7 +270,7 @@ impl CharRefTokenizer {
270270
fn do_named<Sink: TokenSink>(
271271
&mut self,
272272
tokenizer: &mut Tokenizer<Sink>,
273-
input: & BufferQueue,
273+
input: &BufferQueue,
274274
) -> Status {
275275
// peek + discard skips over newline normalization, therefore making it easier to
276276
// un-consume
@@ -304,14 +304,14 @@ impl CharRefTokenizer {
304304
tokenizer.emit_error(msg);
305305
}
306306

307-
fn unconsume_name(&mut self, input: & BufferQueue) {
307+
fn unconsume_name(&mut self, input: &BufferQueue) {
308308
input.push_front(self.name_buf_opt.take().unwrap());
309309
}
310310

311311
fn finish_named<Sink: TokenSink>(
312312
&mut self,
313313
tokenizer: &mut Tokenizer<Sink>,
314-
input: & BufferQueue,
314+
input: &BufferQueue,
315315
end_char: Option<char>,
316316
) -> Status {
317317
match self.name_match {
@@ -395,7 +395,7 @@ impl CharRefTokenizer {
395395
fn do_bogus_name<Sink: TokenSink>(
396396
&mut self,
397397
tokenizer: &mut Tokenizer<Sink>,
398-
input: & BufferQueue,
398+
input: &BufferQueue,
399399
) -> Status {
400400
// peek + discard skips over newline normalization, therefore making it easier to
401401
// un-consume
@@ -414,7 +414,7 @@ impl CharRefTokenizer {
414414
pub(super) fn end_of_file<Sink: TokenSink>(
415415
&mut self,
416416
tokenizer: &mut Tokenizer<Sink>,
417-
input: & BufferQueue,
417+
input: &BufferQueue,
418418
) {
419419
while self.result.is_none() {
420420
match self.state {

html5ever/src/tokenizer/mod.rs

+9-9
Original file line numberDiff line numberDiff line change
@@ -206,7 +206,7 @@ impl<Sink: TokenSink> Tokenizer<Sink> {
206206
}
207207

208208
/// Feed an input string into the tokenizer.
209-
pub fn feed(&mut self, input: & BufferQueue) -> TokenizerResult<Sink::Handle> {
209+
pub fn feed(&mut self, input: &BufferQueue) -> TokenizerResult<Sink::Handle> {
210210
if input.is_empty() {
211211
return TokenizerResult::Done;
212212
}
@@ -248,7 +248,7 @@ impl<Sink: TokenSink> Tokenizer<Sink> {
248248
//§ preprocessing-the-input-stream
249249
// Get the next input character, which might be the character
250250
// 'c' that we already consumed from the buffers.
251-
fn get_preprocessed_char(&mut self, mut c: char, input: & BufferQueue) -> Option<char> {
251+
fn get_preprocessed_char(&mut self, mut c: char, input: &BufferQueue) -> Option<char> {
252252
if self.ignore_lf {
253253
self.ignore_lf = false;
254254
if c == '\n' {
@@ -283,7 +283,7 @@ impl<Sink: TokenSink> Tokenizer<Sink> {
283283

284284
//§ tokenization
285285
// Get the next input character, if one is available.
286-
fn get_char(&mut self, input: & BufferQueue) -> Option<char> {
286+
fn get_char(&mut self, input: &BufferQueue) -> Option<char> {
287287
if self.reconsume {
288288
self.reconsume = false;
289289
Some(self.current_char)
@@ -294,7 +294,7 @@ impl<Sink: TokenSink> Tokenizer<Sink> {
294294
}
295295
}
296296

297-
fn pop_except_from(&mut self, input: & BufferQueue, set: SmallCharSet) -> Option<SetResult> {
297+
fn pop_except_from(&mut self, input: &BufferQueue, set: SmallCharSet) -> Option<SetResult> {
298298
// Bail to the slow path for various corner cases.
299299
// This means that `FromSet` can contain characters not in the set!
300300
// It shouldn't matter because the fallback `FromSet` case should
@@ -321,7 +321,7 @@ impl<Sink: TokenSink> Tokenizer<Sink> {
321321
// NB: this doesn't set the current input character.
322322
fn eat(
323323
&mut self,
324-
input: & BufferQueue,
324+
input: &BufferQueue,
325325
pat: &str,
326326
eq: fn(&u8, &u8) -> bool,
327327
) -> Option<bool> {
@@ -346,7 +346,7 @@ impl<Sink: TokenSink> Tokenizer<Sink> {
346346
}
347347

348348
/// Run the state machine for as long as we can.
349-
fn run(&mut self, input: & BufferQueue) -> TokenizerResult<Sink::Handle> {
349+
fn run(&mut self, input: &BufferQueue) -> TokenizerResult<Sink::Handle> {
350350
if self.opts.profile {
351351
loop {
352352
let state = self.state;
@@ -569,7 +569,7 @@ impl<Sink: TokenSink> Tokenizer<Sink> {
569569
}
570570
}
571571

572-
fn discard_char(&mut self, input: & BufferQueue) {
572+
fn discard_char(&mut self, input: &BufferQueue) {
573573
// peek() deals in un-processed characters (no newline normalization), while get_char()
574574
// does.
575575
//
@@ -698,7 +698,7 @@ impl<Sink: TokenSink> Tokenizer<Sink> {
698698
// Return true if we should be immediately re-invoked
699699
// (this just simplifies control flow vs. break / continue).
700700
#[allow(clippy::never_loop)]
701-
fn step(&mut self, input: & BufferQueue) -> ProcessResult<Sink::Handle> {
701+
fn step(&mut self, input: &BufferQueue) -> ProcessResult<Sink::Handle> {
702702
if self.char_ref_tokenizer.is_some() {
703703
return self.step_char_ref_tokenizer(input);
704704
}
@@ -1384,7 +1384,7 @@ impl<Sink: TokenSink> Tokenizer<Sink> {
13841384
}
13851385
}
13861386

1387-
fn step_char_ref_tokenizer(&mut self, input: & BufferQueue) -> ProcessResult<Sink::Handle> {
1387+
fn step_char_ref_tokenizer(&mut self, input: &BufferQueue) -> ProcessResult<Sink::Handle> {
13881388
// FIXME HACK: Take and replace the tokenizer so we don't
13891389
// double-mut-borrow self. This is why it's boxed.
13901390
let mut tok = self.char_ref_tokenizer.take().unwrap();

xml5ever/src/tokenizer/char_ref/mod.rs

+11-11
Original file line numberDiff line numberDiff line change
@@ -116,7 +116,7 @@ impl CharRefTokenizer {
116116
pub fn step<Sink: TokenSink>(
117117
&mut self,
118118
tokenizer: &mut XmlTokenizer<Sink>,
119-
input: & BufferQueue,
119+
input: &BufferQueue,
120120
) -> Status {
121121
if self.result.is_some() {
122122
return Done;
@@ -136,7 +136,7 @@ impl CharRefTokenizer {
136136
fn do_begin<Sink: TokenSink>(
137137
&mut self,
138138
tokenizer: &mut XmlTokenizer<Sink>,
139-
input: & BufferQueue,
139+
input: &BufferQueue,
140140
) -> Status {
141141
match unwrap_or_return!(tokenizer.peek(input), Stuck) {
142142
'\t' | '\n' | '\x0C' | ' ' | '<' | '&' => self.finish_none(),
@@ -159,7 +159,7 @@ impl CharRefTokenizer {
159159
fn do_octothorpe<Sink: TokenSink>(
160160
&mut self,
161161
tokenizer: &mut XmlTokenizer<Sink>,
162-
input: & BufferQueue,
162+
input: &BufferQueue,
163163
) -> Status {
164164
let c = unwrap_or_return!(tokenizer.peek(input), Stuck);
165165
match c {
@@ -181,7 +181,7 @@ impl CharRefTokenizer {
181181
&mut self,
182182
tokenizer: &mut XmlTokenizer<Sink>,
183183
base: u32,
184-
input: & BufferQueue,
184+
input: &BufferQueue,
185185
) -> Status {
186186
let c = unwrap_or_return!(tokenizer.peek(input), Stuck);
187187
match c.to_digit(base) {
@@ -210,7 +210,7 @@ impl CharRefTokenizer {
210210
fn do_numeric_semicolon<Sink: TokenSink>(
211211
&mut self,
212212
tokenizer: &mut XmlTokenizer<Sink>,
213-
input: & BufferQueue,
213+
input: &BufferQueue,
214214
) -> Status {
215215
match unwrap_or_return!(tokenizer.peek(input), Stuck) {
216216
';' => tokenizer.discard_char(input),
@@ -224,7 +224,7 @@ impl CharRefTokenizer {
224224
fn unconsume_numeric<Sink: TokenSink>(
225225
&mut self,
226226
tokenizer: &mut XmlTokenizer<Sink>,
227-
input: & BufferQueue,
227+
input: &BufferQueue,
228228
) -> Status {
229229
let mut unconsume = StrTendril::from_char('#');
230230
if let Some(c) = self.hex_marker {
@@ -273,7 +273,7 @@ impl CharRefTokenizer {
273273
fn do_named<Sink: TokenSink>(
274274
&mut self,
275275
tokenizer: &mut XmlTokenizer<Sink>,
276-
input: & BufferQueue,
276+
input: &BufferQueue,
277277
) -> Status {
278278
let c = unwrap_or_return!(tokenizer.get_char(input), Stuck);
279279
self.name_buf_mut().push_char(c);
@@ -307,7 +307,7 @@ impl CharRefTokenizer {
307307
fn unconsume_name<Sink: TokenSink>(
308308
&mut self,
309309
tokenizer: &mut XmlTokenizer<Sink>,
310-
input: & BufferQueue,
310+
input: &BufferQueue,
311311
) {
312312
tokenizer.unconsume(input, self.name_buf_opt.take().unwrap());
313313
}
@@ -316,7 +316,7 @@ impl CharRefTokenizer {
316316
&mut self,
317317
tokenizer: &mut XmlTokenizer<Sink>,
318318
end_char: Option<char>,
319-
input: & BufferQueue,
319+
input: &BufferQueue,
320320
) -> Status {
321321
match self.name_match {
322322
None => {
@@ -404,7 +404,7 @@ impl CharRefTokenizer {
404404
fn do_bogus_name<Sink: TokenSink>(
405405
&mut self,
406406
tokenizer: &mut XmlTokenizer<Sink>,
407-
input: & BufferQueue,
407+
input: &BufferQueue,
408408
) -> Status {
409409
let c = unwrap_or_return!(tokenizer.get_char(input), Stuck);
410410
self.name_buf_mut().push_char(c);
@@ -420,7 +420,7 @@ impl CharRefTokenizer {
420420
pub fn end_of_file<Sink: TokenSink>(
421421
&mut self,
422422
tokenizer: &mut XmlTokenizer<Sink>,
423-
input: & BufferQueue,
423+
input: &BufferQueue,
424424
) {
425425
while self.result.is_none() {
426426
match self.state {

xml5ever/src/tokenizer/mod.rs

+11-11
Original file line numberDiff line numberDiff line change
@@ -200,7 +200,7 @@ impl<Sink: TokenSink> XmlTokenizer<Sink> {
200200
}
201201

202202
/// Feed an input string into the tokenizer.
203-
pub fn feed(&mut self, input: & BufferQueue) {
203+
pub fn feed(&mut self, input: &BufferQueue) {
204204
if input.is_empty() {
205205
return;
206206
}
@@ -229,7 +229,7 @@ impl<Sink: TokenSink> XmlTokenizer<Sink> {
229229

230230
// Get the next input character, which might be the character
231231
// 'c' that we already consumed from the buffers.
232-
fn get_preprocessed_char(&mut self, mut c: char, input: & BufferQueue) -> Option<char> {
232+
fn get_preprocessed_char(&mut self, mut c: char, input: &BufferQueue) -> Option<char> {
233233
if self.ignore_lf {
234234
self.ignore_lf = false;
235235
if c == '\n' {
@@ -274,7 +274,7 @@ impl<Sink: TokenSink> XmlTokenizer<Sink> {
274274
self.emit_error(msg);
275275
}
276276

277-
fn pop_except_from(&mut self, input: & BufferQueue, set: SmallCharSet) -> Option<SetResult> {
277+
fn pop_except_from(&mut self, input: &BufferQueue, set: SmallCharSet) -> Option<SetResult> {
278278
// Bail to the slow path for various corner cases.
279279
// This means that `FromSet` can contain characters not in the set!
280280
// It shouldn't matter because the fallback `FromSet` case should
@@ -300,7 +300,7 @@ impl<Sink: TokenSink> XmlTokenizer<Sink> {
300300
//
301301
// NB: this doesn't do input stream preprocessing or set the current input
302302
// character.
303-
fn eat(&mut self, input: & BufferQueue, pat: &str) -> Option<bool> {
303+
fn eat(&mut self, input: &BufferQueue, pat: &str) -> Option<bool> {
304304
input.push_front(replace(&mut self.temp_buf, StrTendril::new()));
305305
match input.eat(pat, u8::eq_ignore_ascii_case) {
306306
None if self.at_eof => Some(false),
@@ -315,7 +315,7 @@ impl<Sink: TokenSink> XmlTokenizer<Sink> {
315315
}
316316

317317
/// Run the state machine for as long as we can.
318-
pub fn run(&mut self, input: & BufferQueue) {
318+
pub fn run(&mut self, input: &BufferQueue) {
319319
if self.opts.profile {
320320
loop {
321321
let state = self.state;
@@ -344,7 +344,7 @@ impl<Sink: TokenSink> XmlTokenizer<Sink> {
344344

345345
//§ tokenization
346346
// Get the next input character, if one is available.
347-
fn get_char(&mut self, input: & BufferQueue) -> Option<char> {
347+
fn get_char(&mut self, input: &BufferQueue) -> Option<char> {
348348
if self.reconsume {
349349
self.reconsume = false;
350350
Some(self.current_char)
@@ -497,20 +497,20 @@ impl<Sink: TokenSink> XmlTokenizer<Sink> {
497497
}
498498
}
499499

500-
fn peek(&mut self, input: & BufferQueue) -> Option<char> {
500+
fn peek(&mut self, input: &BufferQueue) -> Option<char> {
501501
if self.reconsume {
502502
Some(self.current_char)
503503
} else {
504504
input.peek()
505505
}
506506
}
507507

508-
fn discard_char(&mut self, input: & BufferQueue) {
508+
fn discard_char(&mut self, input: &BufferQueue) {
509509
let c = self.get_char(input);
510510
assert!(c.is_some());
511511
}
512512

513-
fn unconsume(&mut self, input: & BufferQueue, buf: StrTendril) {
513+
fn unconsume(&mut self, input: &BufferQueue, buf: StrTendril) {
514514
input.push_front(buf);
515515
}
516516
}
@@ -640,7 +640,7 @@ impl<Sink: TokenSink> XmlTokenizer<Sink> {
640640
// Return true if we should be immediately re-invoked
641641
// (this just simplifies control flow vs. break / continue).
642642
#[allow(clippy::never_loop)]
643-
fn step(&mut self, input: & BufferQueue) -> bool {
643+
fn step(&mut self, input: &BufferQueue) -> bool {
644644
if self.char_ref_tokenizer.is_some() {
645645
return self.step_char_ref_tokenizer(input);
646646
}
@@ -1206,7 +1206,7 @@ impl<Sink: TokenSink> XmlTokenizer<Sink> {
12061206
}
12071207
}
12081208

1209-
fn step_char_ref_tokenizer(&mut self, input: & BufferQueue) -> bool {
1209+
fn step_char_ref_tokenizer(&mut self, input: &BufferQueue) -> bool {
12101210
let mut tok = self.char_ref_tokenizer.take().unwrap();
12111211
let outcome = tok.step(self, input);
12121212

0 commit comments

Comments
 (0)