@@ -1770,7 +1770,7 @@ static bool rb_meta_init(struct trace_buffer *buffer, int scratch_size)
17701770 bmeta -> total_size = total_size ;
17711771 bmeta -> buffers_offset = (void * )ptr - (void * )bmeta ;
17721772
1773- /* Zero out the scatch pad */
1773+ /* Zero out the scratch pad */
17741774 memset ((void * )bmeta + sizeof (* bmeta ), 0 , bmeta -> buffers_offset - sizeof (* bmeta ));
17751775
17761776 return false;
@@ -6089,7 +6089,7 @@ static void rb_clear_buffer_page(struct buffer_page *page)
60896089 * id field, and updated via this function.
60906090 *
60916091 * But for a fixed memory mapped buffer, the id is already assigned for
6092- * fixed memory ording in the memory layout and can not be used. Instead
6092+ * fixed memory ordering in the memory layout and can not be used. Instead
60936093 * the index of where the page lies in the memory layout is used.
60946094 *
60956095 * For the normal pages, set the buffer page id with the passed in @id
@@ -7669,7 +7669,7 @@ static __init int test_ringbuffer(void)
76697669 /*
76707670 * Show buffer is enabled before setting rb_test_started.
76717671 * Yes there's a small race window where events could be
7672- * dropped and the thread wont catch it. But when a ring
7672+ * dropped and the thread won't catch it. But when a ring
76737673 * buffer gets enabled, there will always be some kind of
76747674 * delay before other CPUs see it. Thus, we don't care about
76757675 * those dropped events. We care about events dropped after
0 commit comments